var/home/core/zuul-output/0000755000175000017500000000000015145351044014527 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015145355251015477 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000230253715145355076020275 0ustar corecore>ڕikubelet.log_o[;r)Br'o-n(!9%CMc;b[>Ǧ(\XGfL3e>c7M:~*U糿h[.|yoo\n6^xzWc!Kޒ_i3_.gSeq1v:+|^ǿqڝo>߬}g ԏbś nM "^\,V۳z7wg=K7W0W"ŒGǯgu]vGR)$DD D~u;\iX\|U. $ไsύ<83Jp ώI8&xz_W#~[oa0vs68/Jʢ ܚʂ9ss3+aô٥J}s=7FEbп1 FKX1QRQlrTvb)E,s)ɀ;$#LcdHM|J?_kPϨF&%3KpNGIrND}2SRC ss]QzH.ad!rJBe`V +|i}}THW{yy g?s- oζ@e=Vn)h\[.!=A(%UJR*kl|+"NGC_#a7]d]sJg;;>Yp׫,w`ɚGOd$ecw^~7EpQС3>GCS[Y?D?GS awoap+W9f%$P[4D2LG1bЛ]\s΍ic-ܕ4+ަ^,w7[A9/vb֜}>|TX rdTs>RDPhإek-*듌D[5/l2_nH[֫yTNʹ<ws~^B.ǣ''ASGOEȧ`hmsJU # DuT%ZPr_W_ŏPv`9 C|iRj)OCMituuۀ~s t*;o7sp$3nC|]|[>ӸUKޥg9b2oII"9 1"6Dkiid3N^[-הp|A*Z*}QJ0SqAYE0i5P-$̿=_d^"]}Z|)5rC jof'(%*݆^J">AMMQQ؏*NL ߁NPi?$;'#&立q\ >hl%}Р`sMCכAztԝp ,}Nptt%q6& ND lM=ָPZGa(X(2*91n,5/0KN_Ď6)')SĔv}S%xhRe)a@r N0RQ.FkyZ< O)VCRQrC|}nv?R?Q~}9_bu'b߇ٜK;tf*H7(?PЃkLM(]֟-Xصp&NI%`t3Vq=Mb㸵2*3d*mQ%"h+ "f "D(~}moH|E3*46$A'>7aX)󇛠ƾ9U^}KmJ?t 5@հ1hr}=5t;J|dͤ߯R> kH&Y``zG,z҄R K&Nh c{A`O'd1*-B[aL"T 1dȂ0TJ#r)٧4!)'qOכrXMqHe1[7c(+!C[KԹҤ 0q;;x+G'ʐƭ5J; 6M^ CL3EQXy0Hy[``Xm635ӯ,j*X}6$=}0vJ{*.Jw4?؃ E"#1?|ђP? -8%JNIt"`HP!]!V 尛a;i`qCNG?UPԠ"ƎoC!0[r_G{j P>Qwf8*c4˥Ęk(+,«.c%~&^%80=1JgޛIgǽgr&P29LcIIGAɐ`P-\zʡP=]RFZx[|mi ǿ&Gi_owi[BOdG.*)Ym4`-RAJLڈ}D1ykdW׻"/6sJ%%´ƭ*( :B_o3YKocr ][um#?,_?t?}=rQv^sP3.sP1GNsmd_՝=z1Jid % Jwe`40^|ǜd]z dJR-Дxq4lZ,Z[|e 'Ƙ$b2JOh k[b~}!oѶvhu|8Qz<^S-7;k>VO><~%gi ˿7؞1*]h,*aklVIsc7d'q@WEݰLkS :}%Jk3GIĨmIEQ«` RPZ(D2G=>l |fͨ3t'_iMcĚHdo񘊱;md^6%rd9#_z2:Y`&US tDkQ;~EC#{u9]))wF|~(XA PLjy*#etĨB$"xㄡʪMc~)j 1駭~բ>XiN .E轋RQ'Vt3,F3,#Y3,kJ3,LhVnKauomˠ_~g,ZByXO&Ksg3["66hŢFD&iQCFd4%h}̗Uɾ?qi&2"C]uo$.`mbmƒVe9f6NŐsLu6fe kىKR%f"6=rw^)'Hz }x>1yFX09'A%bDb0CPvw.T/ia v[|mꧽj -ݛrbH9t ֥S`Mh$욕v`;VI&^ϊх{211 VSxSew=?wW&K{Ll)HClba1PIFĀ":tu^}.&R*!^pHPQuSV$.7KMb.:DK>WtWǭKz4@Va3"a`R@gbu!_J5Ґ 3?lm$K/$s_. WM]̍"W%`lO2-"ew@E=A+'h=TԫeVިO? )-1 8/%\hC(:=4< ,RmDRWfRoUJy ŗ-ܲ(4k%הrΒ]rύW -e]hx&gs7,6BxzxօoFMA['҉F=NGD4sTq1HPld=Q,DQ IJipqc2*;/!~x]y7D7@u邗`unn_ư-a9t_/.9tTo]r8-X{TMYtt =0AMUk}G9^UA,;Tt,"Dxl DfA\w; &`Ͱ٢x'H/jh7hM=~ ֟y[dI~fHIqC۶1Ik\)3 5Ķ']?SؠC"j_6Ÿ9؎]TTjm\D^x6ANbC ]tVUKe$,\ܺI `Qز@UӬ@B {~6caR!=A>\+܁<lW Gϸ}^w'̅dk  C 7fbU{3Se[s %'!?xL 2ڲ]>i+m^CM&WTj7ȗE!NC6P}H`k(FUM gul)b ;2n6'k}ˍ[`-fYX_pL +1wu(#'3"fxsuҮױdy.0]?ݽb+ uV4}rdM$ѢIA$;~Lvigu+]NC5ÿ nNჶT@~ܥ 7-mU,\rXmQALglNʆ P7k%v>"WCyVtnV K`pC?fE?~fjBwU&'ᚡilRї`m] leu]+?T4v\% ;qF0qV(]pP4W =d#t ru\M{Nj.~27)p|Vn60֭l$4԰vg`i{ 6uwŇctyX{>GXg&[ņzP8_ n` *3UP0Sp8:>m(Zx ,c|!0=0{ P*27ެT|A_mnZ7sDbyT'77J6:ѩ> EKud^5+mn(fnc.^xt4gD638L"!}LpInTeD_1ZrbkI%8zPU:LNTPlI&N:o&2BVb+uxZ`v?7"I8hp A&?a(8E-DHa%LMg2:-ŷX(ǒ>,ݵ𴛾é5Zٵ]z"]òƓVgzEY9[Nj_vZ :jJ2^b_ F w#X6Sho禮<u8.H#',c@V8 iRX &4ڻ8zݽ.7jhvQ:H0Np: qfՋ40oW&&ף \9ys8;ӷL:@۬˨vvn/sc}2N1DDa(kx.L(f"-Da +iP^]OrwY~fwA#ٔ!:*땽Zp!{g4څZtu\1!ѨW(7qZcpL)ύ-G~^rFD+"?_h)yh=x>5ܙQ~O_e琇HBzI7*-Oi* VšPȰһ8hBőa^mX%SHR Fp)$J7A3&ojp/68uK͌iΙINmq&} O L-\ n4f/uc:7k]4p8wWLeUc.)#/udoz$} _3V6UݎvxyRC%ƚq5Щ/ۅw* CVo-1딆~ZYfJ"ou1ϵ5E bQ2mOΏ+w_eaxxOq:ym\q!<'J[FJ,4N:=6. +;$v6"I7%#CLTLyi{+ɠ^^fRa6ܮIN ޖ:DMz'rx#~w7U6=S0+ň+[Miw(W6 ]6ȧyԋ4ԙ./_A9B_-Z\PM `iĸ&^Ut (6{\٢K 5XGU/m >6JXa5FA@ q}4BooRe&#c5t'B6Ni/~?aX9QR5'%9hb,dsPn2Y??N M<0YaXJ)?ѧ| ;&kEYhjo?BOy)O˧?GϧmI C6HJ{jc kkA ~u?u7<?gd iAe1YB siҷ,vm}S|z(N%Wг5=08`S*՟݃*־%NǸ*kb05 V8[l?W]^@G:{N-i bɵFWǙ*+Ss*iނL?{ƱPs(9vb$Η} % 'ןYI ǧPOUCEm<}:H(^jJ=䄸-m!AdEږG)շj#v;#y/hbv BO Iߒ {I7!UՆGIl HƗHg1$ʇȠh#CBΤ{sTQ{%w)7@y1K^ ].Y$46[B-3%OONw8d`Q4d$x0t8@t]y1T\YAidtxBG:pɨyeNg4n]M؞ e}Wn6׳i~'ہZ*FU{fXڃP'Hl4 ,ŸqMHDCYZz Qnz܁$Jp04ȴIL΃.0FiO-qy䶬j顒x.Ƞ"m@Hy_I )j|s#RGI!dTKL&4K>#stV \'xMgaSZNg8>e!^f%cYr]qs:"̊;isXa]d+"v=x7p.fZCg_Ys;pE&\U}ܫSh])qKYAـhΔ`UаQ/ .D%ES*;OLRX[vDb:7a}YF30H #iSpڮޓJ|̖D?:3mhW=rOf'/wѹ8BS8]`;=?Ųm gp8펠ˋֆ:1IC8qٞ\vXçL ]X/r}7O}Wh,h ;RQ=]u10yiC۔I^3!?H6iUV$aZĄH(CR9m*ǁxFd q w帖k8UKxH Hx?tyteIGbIqܢcZW {jfѐ6 QڣPt[:GfCN ILhbB.*IH7xʹǙMVA*J'W)@9 Ѷ6jىY* 85{pMX+]o$h{KrҎl 5sÁbNW\: "HK<bdYL_Dd)VpA@A i"j<鮗 qwc&dc,m+-!sFɸv'IJ-tH{"KFnLRH+H6Er$igsϦ>QKwҰ]Mfj8dqW+"/fC Q`B 6כy^SL[bJgW^;zA6hrH#< 1= F8) 򃟤lŏd7>WKĉ~b2KQdk6՛tgYͼ#$eooԦ=#&d.09DHN>AK|s:.HDŽ">#%zNEt"tLvfkB|rN`)81 &ӭsēj\4iO,H̎<ߥ諵z/f]v2 0t[U;;+8&b=zwɓJ``FiQg9XʐoHKFϗ;gQZg܉?^HtpR:4Si גt'MH6@w钐֋H{xYEE>6nOj5~IJ|~!yKڮ2 h 3x}~ے4WYr9Ts] AA$ұ}21;qbUwRK #}u'tLi'^Y&,mCM)eu㠥Ѻ\a}1:V1zMzT}R,IA e<%!vĉq|?mtB|C ?dXuWLGml?*uTC̶V`FVY>ECmDnG+UaKtȃbeb筃kݴO~fފ⊈ 8MK?:mM;ߵoz+O~e3݌ƺ(ܸf)*gCQE*pp^~x܃`U'A~E90t~8-2S󹞙nk76s&"mgVKA: X>7QQ-CDC'| #]Y1E-$nP4N0#C'dvܸȯ.vIH"ŐR ;@~{>Kv{) 9AG ćͩ$.!б~N8i"1KФ\L7/,U@.ڮO?mُa ې!rGHwɚ YZgދ8H KV,XHS4OAH$,mr달v9dB.bq:__xW|1=6 R3y^ E#LB ZaZd1,]ןkznxt[|d7W8?XfNѕ|O'  Qk^K|ooz-3,\ϦRY_TBZyX-Ǐ_whaE 98 (oѢ/Р΅ 7ցl6618ł_1/=fu).s¯?nS[{'b{}IStgq$<$ilb)n&  $uT{wD]2cM(%YjDktByxVlᷳ1~jpd1O9Á%. er@3?]^ʧM <mBɃ|.a]]i-*ػWeiبb@66%A,՘|bOc 7O'T*|h^U{ãeeH7nz#<7=li$2cPs6D>TE-&I5&lҬt񗽐eYpCŋ݃k-$pxbڲ&6*9mg>{rtD)wQ`pkKyt1?[ˋڄ5NhfӛŮ Qu8Y4ݯ^+՗~te\g8zHT%{,#/0'NZ׽[q^ė2ymGqF8RŔ.MMWrO»HzC7ݴLLƓxxi2mW4.@`tF)Ċ;@@twml"Ms>\΋"?|NKfֱn !s׭dֲcUh=Ɩ9b&2} -/f;M.~dhÓ5¨LIa6PnzɗBQiG'CXt!*<0U-(qc;}*CiKe@p&Em&x!i6ٱ˭K& FCfJ9%ٕQ·BD-]R1#]TROr}S [;Zcq6xMY 6seAU9c>Xf~TTX)QӅtӚe~=WtX-sJb?U'3X7J4l+Cj%LPFxŰAVG Y%.9Vnd8? ǫjU3k%E)OD:"Ϳ%E)=}l/'O"Q_4ILAٍKK7'lWQVm0c:%UEhZ].1lcazn2ͦ_DQP/2 re%_bR~r9_7*vrv |S.Z!rV%¢EN$i^B^rX؆ z1ǡXtiK`uk&LO./!Z&p:ˏ!_B{{s1>"=b'K=}|+: :8au"N@#=Ugzy]sTv||Aec Xi.gL'—Ʃb4AUqػ< &}BIrwZ\"t%>6ES5oaPqobb,v 2w s1,jX4W->L!NUy*Gݓ KmmlTbc[O`uxOp  |T!|ik3cL_ AvG i\fs$<;uI\XAV{ˍlJsŅjЙNhwfG8>Vڇg18 O3E*dt:|X`Z)|z&V*"9U_R=Wd<)tc(߯)Y]g5>.1C( .K3g&_P9&`|8|Ldl?6o AMҪ1EzyNAtRuxyn\]q_ߍ&zk.)Eu{_rjuWݚ;*6mMq!R{QWR=oVbmyanUn.Uqsy.?W8 r[zW*8nؿ[;vmcoW]"U;gm>?Z֒Z6`!2XY]-Zcp˿˘ɲ}MV<в~!?YXV+lx)RRfb-I7p)3XɯEr^,bfbKJ'@hX><[@ ,&,]$*բk-Yv5 '1T9!(*t 0'b@񲱥-kc6VnR0h& 0Z|ђ8 CGV[4xIIWN?Yt>lf@ Vi`D~ڇŁQLLkY <ZPKoma_u` !>Z;3F\dEB n+0Z ?&s{ 6(E|<ޭLk1Yn(F!%sx]>CTl9"و5 |ݹր|/#.w0ޒx"khD?O`-9C| &8֨O8VH5uH)28 Ǿ-R9~ +#e;U6]aD6Xzqd5y n';)VKL]O@b OIAG Lmc 2;\d˽$Mu>WmCEQuabAJ;`uy-u.M>9VsWٔo RS`S#m8k;(WAXq 8@+S@+' 8U˜z+ZU;=eTtX->9U-q .AV/|\ǔ%&$]1YINJ2]:a0OWvI.O6xMY0/M$ *s5x{gsəL3{$)ՆbG(}1wt!wVf;I&Xi43غgR 6 ݩJ$)}Ta@ nS*X#r#v6*;WJ-_@q.+?DK១btMp1 1Gȩ f,M`,Lr6E} m"8_SK$_#O;V 7=xLOu-ȹ2NKLjp*: 'SasyrFrcC0 ѱ LKV:U} -:U8t[=EAV$=i[mhm"roe5jqf$i>;V0eOޞ4ccc2J1TN.7q;"sդSP) 0v3-)-ٕAg"pZ: "ka+n!e߮lɹL V3Os\ဝ+A= 2䣔AzG\ ` \vc"Kj61O Px"3Pc /' PW*3GX liWv-6W&)cX |]O;C%8@*Z1%8Gk@5^NtY"Fbi8D'+_1&1 7U^k6v읨gQ`LRx+I&s5Www` q:cdʰ H`X;"}B=-/M~C>''1R[sdJm RD3Q{)bJatdq>*Ct/GǍ-`2:u)"\**dPdvc& HwMlF@a5`+F>ΰ-q>0*s%Q)L>$ćYV\dsEGز/:ٕycZtO 2ze31cDB/eWy!A/V4cbpWaPBIpqS<(lȣ'3K?e Z?ڠ8VSZM}pnqL f2D?mzq*a[~;DY〩b𻾋-]f8dBմVs6傊zF"daeY(R+q%sor|.v\sfa:TX%;3Xl= \k>kqBbB;t@/Cԍ)Ga[ r=nl-w/38ѮI*/=2!j\FW+[3=`BZWX Zd>t*Uǖ\*Fu6Y3[yBPj|LcwaIuR;uݷ㺾|47ߍeys=.EinE% 1zY\+͕߬VͭW_겼cazyU1wOw)Ǽn@6 |lk'Z|VZpsqL5 څB}>u)^v~,󿴝} 3+m𢛲Pz_Sp2auQAP*tLnIXA6L7 8UgKdT)*7>p{Pgi-b)>U6IXabPde Ӽ8Ģ8GɄnb'G ֤Mcv4?>HC78NE@UMc8>`TvZ:}O K?m\P"*Wdr]7N;r(qW )YrlCV3b%$zl@'Fĉ$i*γ]Oȡ*L&9Hp6E+CU|JU B HP" xGEWrsWڡ)8TEd"HZ'pQ2hʪ^n6fYr< <$CUzq*SZGepJ1,T@V3ġmX.HU͔ZE(_OM&ͱ@U]A:G82NR99ODpAE5=x>L!pgBS>.}s7*x t^Kʰc.)[ tOlɮvG6w6.b$$$xi,z/c\Kc/C3]iQF6eʳI~Qp/kTKe*4#ўuE(, m`lF‰x(Q5e}[kq=byYOgjl^Բ\ׇcJĶ_֏g\8z~Ш FM68LD1A0p~j7&Rdtێ]-P%Yhpc|#iH$0KOo/eSI^>4NI\W$}}h”fwzA de0ww( <"(4FGިKFM!Cc֫w1xl{eZWsezWec${홿Dh&ò>hAX8޴c|w,ϲGmK O˵ug7oď9R2|TO.(aBu%c;Uuxt)hVEzrOwEVWդ%K%Qr9Ko^NFQ^RQel(IפT׉!}KJOs3Cɼ:TgA5s s'\Z;VqwfkGO`HAXͷ xZb"״wpndBmR ?@Q]r漍1P3YYRMww0%THxLVmq;Az&Tͧ(twtޝ]?*I]$j^]Ts\ OBmdg^^'Q `>{CYOOem>ä>:i\JnAk]? Xdx'@F >T+$_'t&끞LD2剘6E6wT^D}ݠ]OK>s9Hlۻ{oT ˳np BAUN+{XTf(<λUQd(ELDHQ1zyiۚ 繼q0*ה!*xvDfdחyE/] ΝU9\J)K'8B8ʦW. w 'Gx쮼>EMk/-}xpcQ|Uf۱9Sw᱆t;7{E= E}e~wiX8.&38$2\"in,ݒ]om4ˠe\j?Xƽ@)k(!VӋ?Dm..&PG.rܔUHfػ.;*B=oHRLR1aD$/J~|tEknp;_~D+ + $i|l,QbK#^IQZmN&%DNԷbbmA7v{āõM擻ϐl4O{K`imklW8UigCKQ2BjC-apPVPW,j L/tuthy# 㰘 V{X4`kL,WpCLC{c?VX6 :C?2SA\*/nj?nO9涄T߶)sq.y.ncHx{H3gIߒ-ꍵ@mOeUf䚩f8d3+=q6U66U>\4FUUïYOjB碍 L\=p%k.G"k9E5lP!-- )nG/3#cy5!T͵ dL@3i.!=k+p9 [ <J펟1Nf AlfyTC@y+pus\MY+enPJPGBFIc$>^ t .x@K˳S:=-a_YKP-z)vԺ PঔF]򬊻c3SvoQȶm\ &zbYc{>YFP7T0u҅}|VfWXO04V,]͸zmu|eq?8Y$|' F[$Y[M)sCJ`89D?ܕ71@i61W?լDc:>[}fV%$\Q"`bwbf}WCe|Nduȳd6O}-pC_xX3)>ǵ-qUx˜]NօhYzތ㣶{ ^U7څ , ߃%XBV|fn*KZeJ!*g s]8C"L33OѥEnHЈbk:.%\⽞a`C2,f2]vU例/p<,nT);卯,ӍMTFĭg;=K2ZpEwq-CH&"Obeҝ({zcer _rJKu^LLJkK1OsHΙWJ=kn,{w?2Q}oQls;nad.]ZEPqw>y ~>O]-xuWݿU ?׋FzOR^[N=DCoڪ{]z )I=uߢ']Ǝڎ?#3PK)a{ÕRD?ad&(׿RM)u}i@gU":) TRCO[L5Ρwk y݋T!ڴ2D3Ptj y)C@59 ؝2Do-2DǰPhH"KV`Hw7=D}zoR>3SPC/&]_ƕIRNTm7̽i d l=yLb Цj+LܣN@gw{3kmV_pl%J@P6f7'E Kr+yE!%Y1-$C硏YrLZP_9Y|;߃oO KK]٩ x{?+)-<:&F~ SE-Sq]9~N9bXUsL=Jm.ߓ~5%`*{Q& 'e\8,SŨ=oآlV#xRe N>]VHё0Wj_,t)kzr}XDns?`lUk@~8`E V/z`T 8|5ѧ\|QVF!GB~?_EƧƅ|gwZpE(3)c{}(2\q$~8i#_WI ۧSof!7ٺq蝱^"<1ETnCc 8 }8\hxvH`oJ@D>yhxvH04BM^ z b}p ; M>N{Ep Yx$ 򝞄GԷ+V!oQ2vGp2B(FP\!EuSmN#^ 3Tyj@<['xE iގVAmRrr:{mJ=g6umgkԽ$s*Kl,/d(?zt{=,s,`Pg~шI9Νec%`^/HR7 h8~@614>_C)#}kz~5t\UHzU`!{+z+ŸN܎/GAk3ӷ'NG8L'УksNGWq̵5V|˭APs5tnٯ3\AB\/v?>Ӹ >\yy,'/ F&3X0O# a.5_E 6ِyԋbbƖ]R.LhhWz1[D@v_>X4ԨβP0G@Z}ȿtB@e} v`bӕ΄N6+u" ;{i5QkLU`H"W†%ɳ7A븄F0] `g{"Sn9n+ihEpHlH%m3w|C3KeX\e<hpbmĉ`e"MLE`35i!]/Ԟ,I^e-p]'5ݔYp]Jk`#\r@NrF>׳ٴ Cf0Lucr؞ȃBsKF@b8,Y4%P =ܕY ?ed_{lU\}9n)k<}:N-%rB!'Gw_o);LQ#ֺ q(bC˔NLfN! FUI4s_ilJu#RdNxHFahF MǍ9CX(Q&{0Nnq4 &uLNq,U,,y.0q,F Z'LˎGs8<.Uҝ%ҧ(p1T+7†yt:ڰPeA>U^B }v0\Fix:8.98-8: ?@pDLJ\\?SLgLꗲtNx]'7yP藁mܡLpfCyQAl;toU؟NS}oN;`WQe~e7)Z@/2K ޶_5(Y4q=KD3|$dnEIDΤ @ČDб{hM{8ġav,.O&4M'ls8v4Ha7&zx\Mz6g{J%oaehj Y䗰J쩻= [ Yc3]XЍlK^J/sp=ribpyg>jô_)62i@>mŹzQFZv] n0Zsb!VQt{zNl`}n*ɂ8H9v4EI<=$  %D\MQa&f8q$p9,YқElP8/ꪰUPGA <\YԅަWt UU(pm:ժ N[2O`Lv|w6][;_fkkQ M6Ԟ]%6`a F OS~7ӥ0bOU~[ao.T&|*(nv}JmU'נL/`湕Dg!M&֖A'bԽuE9s^dx[W"M$6TT' &-//aB2t B$nC!ی)}ƠG>wXVx` cft2l bHeϤmALڻj?P{ BeB-u'ٍP癄:[ʗ [ʷ'F(&| BeB-u'ݍP[-mA=nz$ۂPL؂P=b7B3 [/oA=n$_C.6=_ ,ˈ bUw͏3jO^WwyY~9IGՃe6-M#tn#w:HtpTR*wIr+MxaʵHc@;!ҿ olRXülLKgp/K].LTOax g-0-2!l~/UOgġb;4(> Eq Q T]yp j|3F Y:V\T^(N<bq3>͢g_(i.:Z S-ES3FG2RNH Lf[YN_0R.lh[ܚI>O(xH, MoA= i1IRYA9Id J5 PvM[*x?F`,FW 3CޖsSJQmlâ)ze@Z΀yj&p>1NiU/ Zz[?urm~6E8Tx(1cYn O& t}1˺urD'q nEL:0Y^o.08o堑P o?$ey~fp2{x2 [αP}F \=e͂r*g*`-vZ8Ev` tTֻG>*HOi:f5Jy|&Xۈ &7 ;qo+W}̪Kɳdayv}NXm9Ϫ^J .lxP$L)Z`Z6]۲ۘbiI!SQ giWr/UUT(_zJkau-w>͡"JWN,N{-%DpGJVr&HQ@43Se,_8\)sx%Ϛ4txb5PP@?ƴZ|"yTz88:,0,t^XdAÈG&1:j>VpWCK2œSMԡ)4.'ԧr".& Qf E[1%:-i]hAKdXBW;dtMgP\y @Jgq{"Ho CZ }_ V-Vp9ˡH)C1B*G*xLej,mdAL`? uKUO@'3/:V h,QPSD*d>?'x^Ϣ}p]xj>"SN۳]j0yW5ݰWu0Q朗*#>)goրPsbFWo+E|!8/>¡[~T:X%&Aj4po峫yZD2zmT??#ŽjkegzyH&";Dy93[ !q FtLBK-KOZFev'ﲻ1M8شp/YI eI;Ds1O}%:lnLmPhkTHkq^ a @8Qkhb/@s?]5IXFYRWKϴt͋RewX6;SF$c={iX'hUa5 PUel ?wk%s s KE\U]1py+wB>5 6˾`s p?T%+cCY3&,_(.4)dsV $붿Efr!r$qR9,p4LP_Iks<h&Hч6IϛRqVfծ{JPԷkFb_ gwgs.SҶNen|,yuc!kraZ|2B$܉m&`X/GC'H Zcs0?߿/%Z}˄?a~xOͨ_3^zX@rzͿ=<%`Ԃ{*!o4Blnmk+!?5'Mhy[^Rb$Ͽ~.uӖLO+~L}eNja?߷f/fz&Wb7O_&=i6<}%[q /vR: sqGmbo8{1>lIBKL[nq)LVK38vƅ2@dZR%Cp@RɅPM\xziE 즵$;xI3.s> pF]ѩd"M)Is魗TDm.eFKXp) E u2.P1.-*g/Vd$i$-)@m$99[D f"JwR|pߞi!Fi Q  v_@JY#+*dĖqۛwy$F#3oa *1!Y ERܪ꒙Sd^?uQ1HnƋ;eA$7ϟISۂ .p\}Ɇ*oЦV]fYRׯB0" *-7uAP .3 t`ܡCDnV% NLI09'㭨wa.h(bj8WyixiΠUO]P52bLgMHl~_qʵģ^qJp<,8Z4M|:iMGT{}z=p4-*s21N{u1B,L RŇۧM\T2b6 EIF\笞tqYnł9>몕,&pi}I:t>0ƦobL &9-&gP1cbǢuFS|9q}Ԇbv"3oͶtW,>JE҅ ~t|Y8b/q6e߰<~m{>HwM=Je ϙR bR?*ԘNIշfץ#9B&tx#S_sʹL=n0i]gM$uXE죉߼)FX5B$V,{TAVR1tO`Q.DT`G7g:cԡ9MDN;76sM2M%+!B-<2%?EqhD)te&q%L.Ft2M#NRb(5Qe.8nL/ 3R_G*!?~A}ҶsḐёv|?8JORs $YZ:)! yڊCn}H6xZ:1f{#Zr)]Fv/M=kˆ..] ީTGAn/0m>pݵ9\pptw-5<5yԦٛ_Fg]2Ex5ջv&Uc,<%02r#+ XqoPe_!KҎ7: MwOanӇ:moxǫք7EJ$*ZbBJM;꼨x8v=u?5o51˘S0^g>{);Z_vVVE%@mɄ%).8'Յ[{ o ;U\Kie%@. (Ϸ36X1-UjP妜E :9q 0Bw-vtrg(y@y7=~YK;Lo+b~([7k F"/$ û0z 93q*֯ϊH)iOlph7ؙO9]T"V:OiS0%*0ԜE{ $MʗSp|U']x}ݱVǥ3At KŤY0j^"iye+m`.G4wGXlBQ*;j. N` ;ׂ7Xukf(utjuc6N(]ӽ=g٢Kub v%T:ѬVKC'1ZڒnoKy bG¨ƊytH21`k[:fڤjݲ7>cx?ANM4Bek|=phՓ4ݕL{0GOlqBh|aܹTt(DӇx'[PCkó՘kM_LtQyx]>ijijўMoDEχ Tiw? ^ 9ά'I^-e*SQ.:wvs,)FR|)WgC$Rf%wAwLN#*pԿ;0ιGY]q.5NV}_!Jy?̈́]pzQ Oym]][2oڊ^ecvq!E:=8̜7\NϜpq_E%P>{$]}}pAgJހpSE! `S̢"/IѢO]cd^F)0EeRZ>99zD2Y(.w}u]pi47o;!ęIw.8e~Yc!!z~6}p&@)ٛ Rh>t"`!_NOnxhh=Q_0`~ƨhR=F%đk.{!}ǹ@4zџ{砉h Ho<7?-[~CstO/޳& Q_EEe,^: *V61E;mtTt 5]D_3ס0* 1wF`@,{~wf\~k?>L.]T7u'E%B(I;Kg]P\@]2<$e8L:A:If+\.w=>~]>YE'l?E$XЅ`1Д @UI ݼO9B_A LF 蠓8'A+zwB(eQy.tH SlP[z,QfhW4¬6=/ x=E * Kz9FtlF>e6ZjQVo> $(E٤^VKjr:ьF޸ę){Nb8$>tI${YDiP zB%& ++M~me(;T0eue tl^%o<9sqpo2+/8Aw>'9JZ-2FĩR~\d24ӃftӓN2\6"iΏY/w% Kbn8,rѾ)۽rv^}Ʃ?LyQL@p뙊"q9(N\^)312>ǀPd䓮7OGңn$8Zi*F$:뛞/a6ړ8i0Bl&"U7l2z |-bX<(j[\vo5kk>dm`oaQt 0f#0Ev {cm3њo/Χ䤋A,)Ercи0hٽJH! ޿|AR;n9Uz0B5Wy69zaz\gIj݇/u'ǝ+өЎx&pC~k|=(* .kx0w QgM4֒ް NjsFa8Lj[F=N^R&њr޽5yi P'VQ*f=TEo*M[mTi.D4kdBU@D-TAPyb#jHpBo<n7HBrZժjLHa'mHe[-m̸]ˈ-]FtvMZ (-6BpM6hNF[t2ӸUOoPV=K;rAXVNAzIxg?ʳ볛mNUF 6 Sᄰ?L⋯:ՠ%'KTXg!g)EM$!XPע2C1[RH/ fE 0K5"De3".' /ܟ;כdA7u$aHYG֒[D0NP!32c}LFZiQ v]{Rc>C0Վz4,LYp}%toKkBnc M;0Fr`RkGi_#η]ޢ,'FYl0 N Rg!H1}&'!yͼ%;]o-Ba^v8e9#S^hoSXfNssBZ ,*s0] ;ݨ{Am3̊|>O-x2qf:=k6\e%X+NeKfoRE^4Kqw/#M(h}h}"%7=m3)?݇[ ~7ɃL/~}SkiXݗ@Hi-eX{^ZD۱\Q.w= T݀{JdD [b!t_WՅnP7D (@[ۃd=޽*/+n }_bC^7~@? =ܡמҲxI}"T3!9Ct6PhCmeZcfŞu_z#2vpϾݥwd6` ) <;3U$j.rzhq [jpxuuȧ"p㢛[ϣt2*!^ۢR}9gQjp̥afˊ+-!FCREOy⼰.5 ,)syol}'u'Ywǝ6(fp2HB@5g,O4Bƕ?3.xgxfm,Hϟ$NQ@ÍN}Če-LSكw^njX@g!\^Q*ҤbN < /^ nǧ6NT1\EU{e.JLcsX/jOdvw 72GG(\aa9랹.'PDz?Ⱦyd13$V0#ǯӧO 7Ǧ)A 1JYagJ@(`1ډf5~g 9~}]Ym:!99|5}N7$7t_Q/s> Pf7j,+RlǝBF >C5bcFMoؙ&3$DS-5!;р!7k(,,8 tE(-v>EZ'.$;.~ޅ7`\;z%25QBk@ 9뚢X) m-P0ߎQp2R4QǬ.`cQ<*$ftsTq(MNۢaaܟn4ZoХZ1ˊc}bx18zeW!4wzv\OMػ)!͈b@>j>3~%s1zmW Έ.Oj` 'yψ.0[mz>+Sr}f'y&,`X<vVrAI_WI|I]X5S3 Fͪd5}S6-<wտ U%w<uV,*ح+v ~Ȅ1M!)ҒϭY8PJCXf2/r⹸ٗ3Z9FfAm9'[i`)d"^YgV-|Mc6s9;/5 [$`BHkr0ɶe 1R6Nyq'kzkxjJA ofYFEɬ @jRNU:@J9ixzƥ7'̤ JHC@xvpR eLbYI!dTk"!k-X{L,4UP2*!®E,\N8/veȘ {!d{&S贔;lqׄ xN)dž͹GIH.3k&Pk ҵ)&޼pu7gZI%aQF|%i*2ԬPlGIt=|e_cP?7R1fm KsU{+8>-o FnRDai-i- 6@@&1WGD%t0}, ypZ-SY;VyߗcRyX`Z.f.mZ2 kM{Ej!^ g*\"Ԓ X3I~RU[cՔ+/c1B8LO[O{c)(siqOr^: $HI [h9ǨE,bgԺ0>ר1ƪM@/{6 a\(!\n0,8hںȢG< oUSʶlRLq.3]U-IBM鈃P,S O5(P' tU& ZY 8BK H {ćR) ,RȄ,B)&&d*}/xbS&,ʘ#z}YAbP{f* A 1$d,NH'kO(+g LCa z\#U؍V˚'/(.kjh Vh8,MFBkկARRkrѷBJaw{s5J],}N}C)ؖox=c,:,\0lςҊ{uv2|\ž]ҌdxN~ jfHeJ`Wժzu3ĆLg02J2'ࢌDĥ[a4QBDXGj մm$8?5պG;?ٛ˥ΏVok&IS&9#Mb",鐦XG䁈f]E7i{S}fؒlKs7PE{ 4]p8_Q3۽jlI 7i K#vTe)#'t Q <$ܮ|q[II^ɳ^LS3#_YMyD3,3iB$X% "udKhʚF]f+Mw65 U&0)WLKy7m5u*|}\X6I'zĴXG"\rtPuW!x 饋&.C}& ˀwUmq[\6&0Ps \7d]w5k{Wn_C/h9Q-'o{waoY%BvD@^1.Z͵ %M `>`e L-yZ!`ӱ "iV~q2W`M_Hƴ=]GӀ:PcS1viZst@8>*3وj'I5+t.~6۱-a 3#€B3:WB[iZA3WV&X~S-rAf^B.o5\FtbśQp&GRͰָH Lilɞ[qBWCc>Ztz0o(^g$ wǚwTRwIOW-5bbD̀kCѻҚ5@ݭƑZ˄R#|*/+ |̔&/dV4[GRjB|%Q]b~߽4Z6`aqp ta7|Ce5q.>iqůނhUE KݬxulY6 :v0N;E3?ѓwlf-L%]S+yi$$~(KO-h(Ȯ&0Q5l;wޚ}V>`GkЄuBd[RV-jڂc5ޭvGXa<5jhg3M`Zn:ϫjXЎekhIIu"n]J9<`߈(wq<y*I E‹k .Ûwa/|ꞨS5ĞGHrq}qYa7|_~HZbj ~gޭզs5!`5*{y|7 #|ugp~ZQ[:0=:)CW2Y/Z?-e+Je߾8I翼b8Z]L\9L5"IG'`R;OjXmv^BU,oPLO:`%w`\f)X<$,-s,,W\G@q2Ȁ=p & b+Oytq/9pS1$Š54ԦK.N)nU+~."h'xGŪTQ{`^Q3* /nR%mn*U~Њvo mvҗ /wG6_-Xz_'~ٗ4 ޿]“=}gxyV6<<+!q('ϊ[pF4^&֫$'4oogEUSxRp,gmԘCNt~ܑ&0e_U", vPV]I]sFߠDoLi뼘{yeC Pa-"\zЪ:fj\R7V]Y6tscF!/ 0yxv[]!!z0>sJ%h׳q>oh*gK)@͎!&xy>d9ȄohsxXq(?Lhx2K8=mkuPE|W(6A%QJ,?pOw8A'Ⱥg" c]Dm8Ͽi@ZxXHp| ( zwLǓ\J)U)@+MO|/AU޿y=.+ *SjLI);:{1A#^UOrWwmze4g혋Uk6`.||eĽlgk`]4,<˳"X\86J` rK;ZeqUOxOXBDQ3vLΩ0ϋ)(<x-xZ鏠^R:,aӠX0)7LF_vi͢ 7prOǿ!캓|_KXSɛDO+R?[qҤƴ%`y@8A,quLuAwރ&uh!H)ZGrM*cqMaV7|9DUf/EgF'Si\|a)T[jΈK0Z_I0ONcez5} n(,S^C#c!DH^_I4O2)}` #b bD$k ,pïeɃ?rf/񃿆'3<}0w7d*/4Htmþou{]|j Պij)FxE4M*JáQA֕^Q0u[{̀X~6k3.3bn3MU% Tf^>cxKf_XїAeVNB M40c4Z01EaƠl1ɜ6G-IY+mFBPZq("BYȈQ&LnfYmfvCgYfNQ"!6ela9K'ڙڇn^;Kiv6| IA2/υY$3x ~5,/?ך?+x7 1kϊ65Y$6*Ŀzk$j%؅}VcPCq|M8Y`蘍?ԟ>y$D*)RoLY&.Z{|+-@*b g>c܎q@Z U'p޸ruFowW~~~.P=O?!E"&qG1 IcXX|lyҿn],+[i:Y(_ߚ^uԛr}M.^:.opח} p/|#~p@EWY1)5VL.OhׯlLͿڱ(3;Kv2ѧuߥoV(t_t2cY*Ceݢ۬+Rw3̓F H࿋ˈI2J\J2)cܴ*B&$4d,5i z<̛k8/p,jѕ|*KPoWdt~t7 0WdoqbZ㸑"8|?r8a{pHV,r LRtO;0TbW_f>s6CYbxss:+9l^qQD_5}cϋ%pǕd@5#utS3Ys\%?+}sAWŖr&zޗy7^zYxZfW~@{,O_?~}*= /wTY~\}Z㴧Hc{X9rTI25_k7yar %$T2Pŭun|\oJst뫹|9K (\,~k6bUop`gvam(0soI_ )-2O+ )>ߣ0ljP9=p`4S@׏}'dֈs,"'ZDhp<}h_5HLkB<.hMĒ\TMZ1AbX'-u}C <Չfo{w?<ߧ+AW(4T˾:ۺ4T;({0dl*v7) f,G{ 幦 ۨ`;L$L/rM%\Qe [eES7(Lێ>K3y&].?zFN ӣSZ}F״Y78[{>;8EٝK4 Yj'C2i@wI;v̶Fv:,:-`W-_7_+.ɠ$&)qETE.39{1FÝ`N{,DJ^_t`rܓ*ElRX^0qNϝ͝r(C(hP`p+s|)3)ZCqGH(TP=g`yE8 S+3%,m.ݎX-[D'INyu?.Vdh 6 QW/ϞWє\6f4ǰ栈iIS6 Ϩ a,|ZYbZ]S"~cscnA95kMUg7;[b"u|CB{T{-WlpkۂM.U6^ [ױ[' =r}>a)*E'K7?/liM,o[ϋ?RPCV}|Hy S"cIV#PCіiO?&fR 7)iEMFۤ&z B%H"RvPtD..?ʹPl~#,|dSp:!Bٌs"cuت£R9|5C,"f8]VFWŒ "R1©̳[䥊P9AC*qk"0w}IvGL*_$ XrCFPTL?&p!Wȇ霆1Ssc9ꐱU9Gr@3&gj 6pwq<ǹ:/L.H, D>PtU[CE"[Jw穌Tr 1g,Fh^NKj8/I MGV׋¹휁yO+p/L(.O*r]%ʄ4tت£R9|o3h/xBU4h/(ZzKܘik]r}n9;["zeJ%Ȉ\ >&Eywh2GACUn-’0 ;-xGxlUGeX-rTQ:n<$$6ٍ,FVD"Oyr&p4r2L-xPqHՔ  FuD"rJ;rf5Q'fln' 2i){dd3au2 /Q$@CYB΁mE 2D103*9AϪiͯy\<.q6r_2R7H=쿔U~1ck$ WEutF:z6fh"4Ab?1 ouڰ|k">Q /= ~`M'@Ξbu7U$9'wZr{ gwBe.\jfL^}q<"ʂyԋ똓i`7Po D uM7fdJtߋݑl]!}))dwlj;wSBe}W!dcؽ͘sBwtveC>U(y!Ǝ9A /q"0\M3yeZb Odyg/-'{ү`0=aZp;qwB\)˸uh`\=H>>lj.7O|!(xl'>ݍ+٢z^*, uF5߄Q9}z!2*g q^/Sv*=0m\T(\L./vH9n<rܫ+`SyvXBH{acP]4)̜ `c#sĢB\P.hL2c; @ծ1-}x^xQ82Q,!&cGq!<*~#f|\uv(p\Po2!5M7cW.sygcElcQBxRpFse٩OON1DveޭP $HBE1ȹ!Hs#0CN@'| {@^mbg@Art>^fǡW) WIK'.-D2 0Җ۾hW"6=aׯLNᮛ\ &!8/03hGG,u %V+CGiZ ": ^ۭ/UfIF[>'2w܃J|&W?pcrW =A6AHFVfaC5+;!N?1шkwα"RCF[. "r9]{<(M$R9zgNH)*co?B>Ti{_?B$U]xx5Y@:.ѳ* JK&Lxza]s׃-[ $WCƌSg޲YM3H&SɾJ)M(#8~[gB,?:P2U뺊%T8e'D^B9lO@zꈥׅvruϷ nt/U!cƹ0D=4{\J AN{2(wZ}hu]BxymūMp#=8ZdR^F,vD9tĺ3qAC'{. PS\xT4A A Pm>9'zvȘsާ!eI)1BCGej?)aAsa-R>v&T|V4S@Cd|N CƜqz6X gr|(PBN4cH#ȇ*7'>0C YWH\H@]Bv3;VϿ.PO~t7QA 0)cS[q PLvx-5{>sBqUcRDpܩlbP]2u%f R$][9Cghi$,MBs\6uH"8wԖ.|[CƌEPnȼ*.2*.. 'B:QΖ`!\[7"D8[ӣ1ǻnӭw}_ =zEч\{OHJql$KَRVj`GlW|TVf:ge$ 0aKIٓT1/Б:#Um~QoOϸ(!&K^# A<2$HmWa C2]&gt7"d$B" c1 .!ęZMrfUDi0CV{jBxR^Jy M#܆# cg#-.իmbK" 1@ K@<2X.M݄׀bX'k9ZF2)} Ohw{ fٽxʖ @^e3Y$sVzn# JRJBZAJS͖>[X-yL𥉄k)Ixm-d %la8q#%̪8he*͸F"EC NF};__X'QK Bup{%.X7JVGp4QYY-muICl>kf/:HZ rd{ՠ$F5fS(Pg}(xʀNJV˜=t~g/á4x{Ly[M|y?FSizYe@i"Mzr9hkUҙ$Oxeiˬf2!S[D#xoB )ے6 Y"5-DmpN.YP8,@b׼"R.ffv~F߆vne1RU{0񂠓:n"+愈Λ=:zƨ',3r3 ᥌WoǨ4h{S<WK :kU~MUȹuyd/ hP!ú`IDj / ֶڲaAqɍWh>XY7 ]ty*F*E]C"utSYȬCJeG 6|RLUK*KayV A!bA(e>P3̱o72S*aHz)kAhC4+duKg_ۿMo˺Y-Xˬog;f4Fجd87*5ܰ*¬VSb9×hV UjH)ڬY5+ "xL \ P!+B&aN;a4f?tx;*+$ZmɴT*!;t dڤ]]u;.ؔ`'QFV::xWJ0dpeOcc_X["+3~V7m,r*oGƁ lK QG Yrj̵E Ѫ47鲽 t(R\s+/XѪIOeBvm;#o1a94a;5PWѠ# hGkgWsstĤy`TNlrI*$ӾDRogmbK_?`؏"+r=]{j\1 <#g |39Q*, Zr2ݨ{QGy+xda)??o^pk'yQZ= FCZЦTNfNow<}]QGT/ץr&Cz5z0ʢ~U zgf~{}1sӡ/8"ɔ>/HLd^s7.yE3k~w y*_,+e%uA6YtƱק' E FR)z: nA _jn빩¿^$IG 1 W,.i810vH;dU`MJ˜Do ?!l5Bg]6sd|V7 SE5 *#FZ ["UAdx$/KQA,)%xuM. sP *V只)dR0>~Q)K C< ɓZv hh;Y?HqcD-\`6tg GXG ,WR6(o ;t?%~2brζCV'j8&cW7yLlmGbpyQ\7(xTb35q @߅{ṒN^߅]6.V 7>XT˜!ϸӾQ˂,dyYJ綜3mUZixf ")N}>{w 7Ѣt&%b/Nn{Ef[|˭r`%\N?Vk*ޖz%lI14۾-"<ٝ\ʥc^RO^7Ѩ4Z֢I >\lMb =L~ vxO~5L Fǵ*X#Om0- OKQi,f l*;Bu"kT*&,du\ɱZ`G\y&dŽ@&$Sf/7B*ӂSXA=hl$xö-Th>gwYt<@vZn() tm3bH+t?xgi>Y{@O|ʹ)TviHf$D{;[>!Cklp[*1"2m5l8qC**%t^׷f=wȼ@ݾm4: *&ˉbmZQ X<$w+ÅAo{ wMԭ_<]S5]m:@mݘn Nq힐KSgZG4ڣNOUܧN!_~Xې22xIHI}K@ Mcp0Llθ:xiCݺu;1T4l _(7$cǴ#_x,XjF ix nm<.BKfᴣBgqyiK̎  /ƻ.RRЩ v1T "g#5dy 6H ;RCf&9=@1Ə=`MC9F^>Qt{3UmEvzVo|li 3&ȺBⲧ8ZWw?0kk "?sݧmq5-*;dO,G/q^.jfLfljV ul~6P .,ц:;nW?CHTa+ZbBuɢm\LmmITAZy൪o%nu`❷q}\dle,ZE428/+8A\[ҭlm s~CH l*̉%>Zr}@=aHN5Ӗ\5/D5oג"GQ+'(~5=;Zssb(ς`χ=<;5 g6u|iD- j6{CYϮ!A -/ ~6X N̴nR'亣nl20Qè ,[l *Jjh=<ި=#Z0$"WedbUv(gc"g"XMGR)qN(v9hMjb5 r5CT^l "QraM9ٟ,|XC L0,8t4QH,ʅfe YimwXCV7L0& s&HRBq(; P2G )7[CճXFF]cɍsncpϯ~ۊ uCCmņ!lEpe+Oh*N!%I1gJ!jLK'8!01hlKUeCnB ] ìKJBa!;܀LrӐa9f%e,ZgjdXƦOnNj*j)D|;^q0;$YIw0r|aq)p> ȕRS~6*x?fErz\ŷ:(aQax*g?7G9WJf('ı??(-Ůup`@";GШ bjb5Rڱf5럿5u.F~1 4-C/]Ŕ)Lb40hV!ԍ7.`D=2p-,AH  < *ԩ~buUQŔ2Lh1}~/uts9Kg7'Ntũ pO @4c&;zU~0/;Gh7P3ؑJIۑVy@w?uN~9>â'(\ƀ9IeAK&)U$H;}Ddž(}7猺*t\SgqbsmC{ 1vgDZb4c!P挿,/ӆ:{"p?C,SL~X@X|'C۔XȮ8u[\ ,M{HKE'⨱щB|E悜x ŵDa"DZ6H B|mZ\p,3pӜh v`tLt?l1Z:ْo3= G{<p'ǘZndɸ؇Ax c9 G8Z^Y/z`?k PQ(T1زfcDZfN(R=& [b.h|J`HkE$W1"yI,еƿ0ey6;RsqJ;zPQKAsՅkUW$&OT I{;m)ΖոX)h{<*վ6|zjh#hmw| j/.q50olV2DQ7iH?RtPG"2UX団KE 1^ 8 D4^ R8J( {dКZ V~& 1B{j9P!z_y?L1u1b`_"nkE)\!nˀ0#F$) ) k$#z; r2 !>1=Rz8OO2ntX{vk.O{b\qY^JOp!gE?,۷@}4{{6W֥v hPT\m% Ij,5mjCzHnn0Q_Ạ&2ߣX|{mU\=۷o&MFf?F߿I?cMuV?,|XǵrYSUa3Sϕ]Tgʎ]6l?46h'O"e^Qk}?7}]Vovy3vo|sҏ2|(]ǥbE4WW"X&CTbݪHM{j5j oF Z͎曟$_o7c95k@>ܞϊ90NR!/2&l B8U*cJNdhՈ8O(s*OTt>53u$|uliw8 }CE-2.TӛɤWoh5Rx쫍FYU򣻫ЀX&pg#G'FOAwO78wPl~Fe]q{%U#mf&\ilkTWl Aq4/|m^/Z{]^k Xl?jhY,g[v ZƉv`~N6U;~b9 ?~uL,fzi"뾜ڌ&H+mV p.{i=>n( Qvj'IfEd .Xa'WwsŮbN{*{\}|b"Qj%R+mcCWU5_^^/0UqTL dTCҋv\݋P;I+ !#R)K\rDb A}zmۜP@]N(T1@Zeh&H,(R^ ) `6(*ц!$*93}L{WAvHR0[\^8J#[Vg8q,3Z$ *a;WѡŧܚO^ƒtahNHB`f hpTKBsV+=DH *%z{S$vـl~>q gX(`0FwFsZH/݇4Æ]B]k6PlH~Yw]LY8/ZM\9"Dq@cs#]% x*׶SH0^!x$Dvs]Feڤ@V$2Vh 崔"v!u❦Q,s2&+'dQ(@\Z)[:*5&IvhNIf'9E 6[e@)J~csp݆*v{ ]ŮǛ_Bb_/!K)bv9_BM>KHxd=WĞxMJ0qwTQ`T KmYݒY0ZZNl򹗐{<0u_~X+Nvoh׹/'qo./ei< T/Qy<:t'd%!%M>\ƒ5in%_ R`(KVv2VQڠk}t_w\ pbOk}BZn/0w>L,) [k|RKo#jv_11>̗Nw'q| U\ԁ55 q*C2>g_8G>=JHBdUaF2<idYI KtC*3q`KCv PKƱ.2#YRiőʐ fj{8_!@C=Kla7_%CYbU)8V= V?> ŝٴo]4݄?]: w{(!ܣ<~J2udZ NE3t<̢9(|T snq}w;߼ L/vo~wC:Z}w y5)YmZ?g%H z ^9->&ý#{,.h@˷!R >IB0y?eKxU Fj& W`<  yiVGep̱Rك5cqH\x0ala^nR4C02hnVaoIڈɭ "[cr텫 UVFZ !xCCӇFkZ*sjg/.woZXӪޘK;siL97]ӣ28X}'+}v28"fs|>*Y+AC62ts` Q˱-TՆJ "&z!+O|M+ {r~˟_6zSF >ĦGGep5 2H踤 L763*iQ<[%FOi8'N6΃e'bPC9Xg(P\N'd0~w-h3?)P00Ktckt g+iiFb?OY\rp֜la;<oF9qp!o3f[R3U$T9,2.<s7W!pFGGep;_UpV;d rjE{tTGHʛ뛴K%s1i,Sۊ_fNt6Wƥb㞿d.`Gv;?{$JQP=wA`-&ZiDG)Q]!gorwQ3'+أ282 pdNaox& ZXM ]"U5lLj*`nGO"MްG.& kk@|ڛq9*,' 4Ʀ3¸@1OIex Zuةf %@A0`b%7ŦGGepTt \cRGeRB !0A&V{^>[o^Q#׬B[Y~q6,9(2*H@Y-7m3!!aQlYK8VN(eO@ЄF` H`w!{y .n4^lH(+\|eHp)H#,Z"J8?|ȫ|.2(60P압A0T[@cL49xqɡ=:*˃C<0`20> )k!qJ({Qܘj- $]?^6ݒ],U XTaqeֳr1knff~V|VS]*@iZ>|-= 5LK.%,626VF%L,  _b+"tݔ1ӁM`)l2:ncJ\֙uTG_: ;E[mp(gc2IJhUʝoLv1p8ѫ9}{tT5C&K Dc0;Q2Ê֏!dlN_N"l'2,B5IYhdg󉶪*T#Ezt Lא}U^hǬHzr"lաLWmmԲ'gq)RD&t(ܟA^KL4@ tL[hĬȯ JDYi[Vf݇w.df! )_y |'GGep Td4d|"21=_Y6Q~=C'BH2 q=2g7hh}sc{{:6q{_si !Yx|s<L[HhUFM36v:7Tuq@_QdžGsLg+R=N!E9 ڗ^}fG>Mf+Q`#Q_&g:IYQVugsBhht鸴 QX'xTqO wإDk+;+ﺮI s6a N UYU@bhlnqYuOW@]E9 Ǜ+Y5Cڣ28BNBT,? Hf* >=^֑_ʬ3skWkr=Ns8Qh nmhBڥI!ʰ87M~5,r@xeD QZtGѴ|j};]VgBTeGijY5Q*U[l= IS`rc궆'\ӢB̟g lLHm?dS([)q'NU%΋x^\%Ր>=:QflV۲: Nv '&ݱsgӬg^ ل>**CC w0U0 TrN#ʵ'0sLu@\eu;UG䨺ĉ;fu8lL!eDYBy|Yzd9Ƥ|,jYDгIgbv6zӺ xJLH pC^>wѺv؜=:*#X鍚µ3ϤWDd`SucIT,7Yo] fgkԚ,MH Y{=:jCȲjM^}۵(!% 'Io<mrXv'q}Ӛeh;jA;%7uW:~qGӵc>YO}IQC)iwrj M0&u: ?Vnüq 5Y3' Ej?ބŸ?ã,歙rAM?wI' dVϯ|Yfy7~C " VﶶB> _Of /t ֶ]40_ϗINvАvuw'La0ŷp,J_s忨Cֵ7?w¾u ;4=LKwfn/R5-{~`٦kueoP0vY bq6HFojRz\TknyH/0|xmX^O݌gn 7Wa*|m|=|dm=0c&E9'|uo?3?L& w y:kWy7/>_}tsgğ7}xm?[mzA>x|006پ/)w:}NŭY^,-[]h{^jlA9 Mddvl]3 >t^fwv y;m? d?x$^Opw7yެ^\`" Ւ -LuGVIŷU>y <7f|m_}=_3[Õ'?&ʼnl/6u"<%DQ<PF4{8S{YO˅ ;jEPD{ ;׎wqIw; %C\Rzg8`zy=f&bjB#QTwTJ 6DW^E.V:d ႉtD4hiB$ Lc,SBPJa^{yB4{ DϨ'Lr@7<潇-S(> !/L#0wL7:u c%Y.(uW/N<{68,(9ŝ_>7r-95bDU1Z,XyERK+GDցr#ww~g+孕R%aeHMt :Ui8V"v4-UEǙAPIjuڛDHD"# Fa#UH M b9κq`uqjSHSwcFiƎjRd FQSpDu_;Qgg)U*E1@0J(cR1Ex/Q;^V1k'&Nܮ?&oPo0o0\h3ʞΧ7GfWcVDfybhbh p^I=vD!E艰N4:q8V]ߜ8mZ,&ȢC/f-*^t=kw~gEj"!P-1nTOT+ɄVz.B> 5}u%3"poe6W_+e">VxˉU(#8 > Ĕ1D.xA5SpRH :< ۜbrG]cufvaÂ^PlseV{˱O Jq>CsYv6K{}:"{ ^ky6 =u`O?@*,C/!o =_4,ޙRl3۔qŒ_|F)y§>d/HmW/]~8 W9v m,Q. FR+iS ƙ*yHHS<Іՙ/1=ISv6yʎ :і߃ww>5U$pԿX+3B^*ʱR5L|S?_u'%fȇ]z€"de.vUn_T?Gٹʪ7,u1Vꗟ?рuG֔՜"/r"Ւ![t6 J$G"2r{EdVf~M~Y_Cr&d0HOs2y;.h:ޤC} F۽9}[oz9yۣy~@GJvYφZF:Zjmϻb@K_\8s .\ j 1+f; .IKBQQ`pj6_PzcF#pA @JD:12FGIX>lp;!4D"r M &u՝M9"MʩC#KPG78za04ċXQ_ڢ)6„nBBb5#T')v1v^h#b iգűPe t\[';&#{AגS6%gL)8UKqKiﭕEO=I7Vtt,o"?u`&Y#R _d]NmU&{2j@72뼗Y:e}|޻Br]%y9ec[ύEaBI$cY'/s4#;)3l-](z =~JPu-A n%ӷPeA7 <~6r؂} % nf Pcs}t1w6^ ^Zu!E_.#^b|P+>/C]o]+MMhM`/_`OA,m.ȶF=3Z) 0H3]132sF^YS0+_`xRe7Cz |JsRxhi"tJDb# 5ъP)K+Rn.VeioO疛' g}7oY>FO[>ssCbB_[_X[_֗+n}[_֗G$9TCYBGCRFu64Ф _ ?zPz(wN D¼:P]R[.RACO@h A+4ƙrK4 m^$'u<ǜ!.Ţ0* +16]8;dzL`d'`п|jo9y{qNQ Nz!\z/%G@u,`wA M8$E.VX4;ͭhipѳSl$HD!ArKc= l>Hϣ5RnZ,W3iEAT T% Q)C8PN:FENYgQ ԬC?>҃NE9J<^D gIO` Hyfr 7E$]1ɿ:@z;D4%N9`!eMH\`0kJ/ 6)P3'8ky=+ '<<ݰ,+o|ϱ.B(CI)LFpԾ٢ #w tt( Mo9$xKs];hf7^I\ϔL) jȚ +6 cEԈhQ5@ ɶjQt eK/ʺ?uLP8%s:PQ)|BPό`"_E'Hs])a5!FB1%<Ȅ`*ƩRΆLɢHT}RJ`ŀMVfu ۽"Ⱦ}h:oW5wA} Κ; 8LwS=NWh5U.?&|籅uTKz; h$`Ϋ&Kz|y Gȗot׽~%b?3h9K,'*#4O :?凲op_ _J3)?E:o3ɒI-h;;L ^F W&q~WD*$QP*ŠdDN.rb'xO&StrW'ojI"jdl$F:@T:. q"{b.JJ *ϗT"H,#8f/Ֆسxڟ4>4'$3 T@βӽE)~LcASķ#|,xq#S_)N>۸:]79ZdiijooE}inǟhq|a̒͌mJT5uyN6'EnT}ȍecO, *2$% #^'Fc,K=w@6GH_R%3 bUTzO eR8!Ni$|Tdp7x~ÝAgÚȡpr?Ue+svf D#u&R}w8)[4Me㦲h/.Fn*J λ2*&('& 5A)Eg,谩rD dl!`P6Yk2>oeǫ7(,],z &ꂱwE_j}4݃4 ۏlIS\w `ŸQL>Y( dR6?{֍ KOn`n\\Y;5d2/;)ֲPiDIα ْrR"Rd|k\DL('G=G'b 2 o ErQMSH4BGFDψOmu0ז/&ƢjΆstxHza9(gcRh!WPsD;ќU#D㯘D&4~i7qk(ztxdhNDh+ !n![pkB 5HՂEabB=)Rtsb=z>cBpCޕ)@' F}\XkUUg^*C{guCnθ'nuO)/ֶYW?M[.um2~bbۙ䀺/oySV(VARQ)@ n␷~PBT5lwTrPNe\L&bvHw]$Y}d]B4ɇl^deJ$;th2m%_ϓnr~d~t~{Vcdk5Y]и9(x)SJR\c$]D>Wa7)Ȥ)HS^iFڱ*[RB?٣`P6{rv^YHTd 3RB!:4%KcQ6DO~V~^=\Ht>%H+цZTZB0)( L:tlQR;۔e+gxwKպuZ{7=l'gmTf tXc"iAFC!+.6I&IxF{ y [H!iTbfXTTb58wMZ9+u.6%z}=[ Zrr2pױ_)O7"^㗻xdl˟xVZ˃Z7ɘ&h҈u糗_'lٹ IzNԳ翽n>ypwH(_+˶hgUlܞĽFzeNI,f}MKdd6_p~opAŒë{Rs&C4R֚q@HR|0ѫc$쯟V̻7xv: +V VRNUB4{#$KMd1z3yᣳ!3!zGr7HBJU UXLɁz':3Z=OZ3;*GɫK-}Wڙs\x ~y-Q[AdHDkE?~.**h>6J8Rرº6c$ȃ2jf[UM(:9)Θ'[A젵5$=6~d *n:4~QMrR;P6#(حb =3z<S9*Q,RΏX[^K,AJ#xJTBZev<1 ^emPK%j\E+Ja~#$omtDY+ b2߼4VŠ !xk >EQdq "Dmie ={3x$ٍhQK{Sh#ȤR940DoH޸#h&; 8,a6 1:'itoWFʀUƖX2{0Ԡ\ =G'*gҘڸ9C%toTVK*2TÒu֠!z׾γ+b(E^䘌*$Iʤt~~y,K%0?sFJ#~\U8g9&>@'"klZY??FBRw6XQImvD8$qR#$oMwzKѹj5W(ІqIFo$ &( B~ jS!Q9p1wHkEl檴G Q^i̊Ѳ2!rr|[25! xz[ƊȔrL+S%$qɯsc$dOp Շfc1#N(Xқf4Zz4md9 #$;AqZԗGG.E{iBVodvY$o 7% 8SIGn6i[ֺH8fOEMŃ2d$:y6C%oc#75Y|' t0J#xouo6Uk*X簕pd1[ bdB\(EYS\lOMSPJ6hVdGU6HO$OR@>@Ƙehcioz3( MV%TE*-úc$tro-߈"Lʚ*r̈́5n #x{ORb ,m45!Oڤ#G,g3FB m e29,aj٠pXBUNC:?FB{/}?۽ZoL8rS\T`*W0+xX, CdMqV `́ 5F᭕2N4~.~ 'c-W~}{xD]JOcb1FwZʾK%{]_burfB\]Qtc|9)k.d%[LxqgMCV oPl _Wd[ T\AeNSqm=_βk9'C} + h+h7i yZÒ/p67&Pr:cTl$Jl* 3} b_! # ڡP-:77tλ|`e p񽕀ëýߘ'/<ޫC^qڏSmdoFUkӚ_kȹ#=ԋT-6%fH0Žg+)xJn* '[.0J b~ZK;#B*HA%*& tp'1Sdʞas$EH |@;vOK5)R]S"AbX@eUk)bWbpLԥ{Pr94Ę b `}kTMvRFԞ=K G247i:v7$AtpV&c%teNjcBR^Y)[UtI*fAە;dI!I (A21;'ا<PtH*QuY2XE(lLgLRmqnl*<}>OESTy*OESTQgUTy*<}>OE_v|]٘C+#{\Cyj/3#102ejrF[*lKJj PPU+4Ԁ*'ا11NA[c9wk.\L³]ҬNXwj?^Gc5vd*TvofPPƒx ;Nd:߽s4 w!l^;:;kוVZu1+Wf||#ϩJpѮa!62[oSx7*˺D<F—ƃ\Xnh"{[`e?8{N/*V[DMlqrT;,Yv(qoV( P=\+PuM,ƒŵ+wŢǰ`QT A;T޵q$e_v귀x76Y`\~ʄ)&)Yb>DI|IF,ꯂ6DB-M\uǢ$с}2M`vūѫ{,3 ×1.l[.WeW:YN+A#Ttp43j4:KYqH{*7V-#߱ '*jkPu=RтR^q} q\9ɕIf[kT@MW|[z-e^!vt]s;|akq-HkOH ЛT/+6^q$SO%|ƕ6LvZ|ajtCmfqtZ؇r6m2Xkm!ffJ;a=ckxrf\)>l(*kciDRrT:+1nj Lh 0W涧5av_=;&A됏B"- 2}^ l>HϢ5Rv,I)` (ȑj\·J$@B':jgugK=kWfk?>' %Y/Y"AA]ho2 j0ЬHI2 7"VTM<$W;Poa,KDSi8X\6!LO0)_ mw@֠wTg%:㷏ϝLjP"@Iԃ&D@:AHF,-1fce"Ȳ{oMDVI!f&NhSVbG vB+.q*exD3uQlmFDoXtУQD[V6^f:0T&9g=5K6ĈW Z> zE=S9}p}(њ푇gHyQU$Bq08$Qz XhL#" 9*r.-kv.\3e"wv֥#\Lʕ$pr:pByB8( MNmҹͅ/n z&1< K+ )Tr8哾(vS=r&|Eޏ Pm{KM N 8 Ml/T5:&'{gFPTSJ'9[X60Isn"~}C|(I`>e7y#d'N7L[0uAqOMF՟4og4>~w<~pXyN9cԨKzq;?]SO?֒D7hw=!Ht5S鸼 w7\DvTN=|1޷9*loX`<lQ4w4ٿ%d,oD_F (Y- _fy8賽/|Oķ&v) B#B,O~9H ndttqµҲ {{#ʸN_OhflU]ǿ[ v! OܨʑeX'A;ɂF Akxb4B mB΂͐ŘvRh2+@fbUeTz@ާ\#ɤwqR$tX)NB[Wx(#{3y۞CIjOReO_ZhBgʊyպ月:ekP81_'L^> .#)rkyE !Oԓ %em{7רR<{End H2.ܪvH ;cf[ }E :?:2` p.\Åkp ¡Ugkp 5\ p.\O*\Åkp 5\p 5\ p.\%ddF6pu[n m-\Ppu[nu[n m-\pu=m{mGxVx/p7[̹ۖt"NG~q]"J^Jȵ&]wm❡5ٛkbNopO4ܞ_?'?4A4v1q3=W^v>N料]x5bܘhۍpBĥR+[byK竚h, {T{F Y&^&ztf3^38 V d}AjuX,[왎3k*UT!䜐,ܳcƬy47wQmxsۏpS =}7%WVt=ϝ:}ӊ47kiEt׶ˎWvԧ "1M/|>^%49fƳN3dse}j~ޠÜ৏^1ܑBfCP:ACP:ACp O=sXkT. 96xהY֬斸!IŸNl] (>pW^n6F 4zfl $ 2Iēf 8.",A)>c񁻂W8ЊA#@P>>3Yo Ĩ|/@oO@];;=Dw?׿lQsLBc /xYfh﮺cKd(/Z_8gZk-[KJ@E#B2~+#4Z2  It@2z LEϵ+,CAkj*i*&hҼ5),DU g,CQֺ96q?S<л\U2FM향̍$j|kmͯPo9SLV*2Fih(&#(sj&4GW-޲k1BJ,+ۓW&dMسNqǏ$s/!2Yz<tۨTP\ B Q^k")9# aN QjT\4)0Rm #ɍ B&d6@mZHL{Jћ4`C82w_\gՓFsBѨf$nte)rN>L(hQjey,I鄌K}&OD[F&ON&{BX]:&W|Nw"7m6Ff턝Z$׮cx\)3ArvϪF;kwӜ%$;$ u)v]{ %]˃A%p`k<Jr%~o^훉{ﮢ@sxkQiڅF(&*R(bxpdcDMp/(L|ADVn~i3+zj]g>N}cMV'n%j c{ik'[]rs4Kt9~ڴg6, 6f*l-WU7Í8Q>56_QCqY1Ԭ7] Oy6*|35G|Vm؉jYP;k6g{8Wi d-}0v=rfb Y 揌 )q+$eGyjR)vv^;~ds&d| _3&VJQߖKk~edYz-t﭅6$:Tʂ۬=rtRokv,5Ex.+;2?i+uewzypA8cI/:k5GҎ (EX/ܜ9ocCf[,^)ߞɔvgڥ2}]ڧJe]ښܰ*a4;A?VE=ڃ*h׋UtSxMmGekefKRP*k Av$s0ƲAR)E#5u>ԁzdlox)yyjg53ycӡE*U-KڝK"9v'ϻ--.qgX]ixt:E ԮPʧh*po{BK]O8첪 4&gu ^,[W{\=_W/euo q_چ}іmRh]glC-D+a1jk=UbA Ŋ%_B+|S28[N_Əi{wK.m>hG lOAᄊ=[ {Wfh 36VtJBB3(oPMlt.֢&R%3´qCko_eQXgX1y0x:nRd&1/oߍ${7YD+<ȳ׿mиwoausop/dYlaD3dYB-Š1ԁ@ǎ[11#0q u7Zy-rFFĔ1Q1F@B3 S8juU10VyV uߍTxGDJNK?$s2JaQiƝV2b@^)<5'rcFGUfF>pr^U*U\b1^~\>&%ѿz0BaKƶdȫɸ-n^D`(xmK&/OE ӦtwXyqMI&ƅo{忥 զ`D˿Z/%Kkj\ezA6w^_/F͗?9>g}[ &$fjb3గ0b􂰿JVE+}]ʳ$'+0uo}@0,o1E3y9<{ev`xg7q|Ym/Vlp=Nvk䆏5r#tۥ 9!Ə\v͹j@ŔB}Zf//ITԉi^٦Juj<(˿~GSuwVޞ^}q (ꃋuwh׸vh~}S-ի]{56v^ PkkPHLHu$GnlFuufyh\qYt?hpXj?A* ԎJQDnu+{!#i65*f|3@eKzCNxg?v.~]տ՛_a.zq7p,`e pi<Mx0/~h[ MchMڤZҳT6I2㮉OQzx۟?EC XY͵Nt1. s1+J̇&_W i,1EA^^; [OS(xV32ƙ<'={z.N0Os'{ #QS' 7M6Y#S N2MEpBBȝ h% FZ/$aX>uZIח x!zbS-6?ȹ].z[X̵Q*g <չȭ`)$d{n 0= }rA;#( '3>wZdk)@2Pl)[eˈéRu @42ZBGl8Si K&"rv+RlJa^iGjd~rV2wE%K@)*2Njm:MtB6" RIawl0Km^R# +~vU1i]7-vlok? c; m+zv =j;tBH.uC\5G#*)3yz'hl8"Я\ myq9?mt,S>e9 Pu #z!U6H>{)#"b1h#2&">ZD Я<9($IP6E9MނfS$̆wu]@~)[6\ `Bh 5(W8s\(rO~Dol/I6(ޜ>轼jLe]f)?߄lR ܆,̃O7ad?OR,ש>ctv}֗{hi%;ͽ&=1=[It>Cd&#lvyd-${?}dzz4v4woSk??NͶ*rFBB,[JL.0"~ u/#Bem`G0n$.d%ЧE*$#opHr(q]uWuwZIm\T$(%?ȧF&" SooHb敒`"O y Bjp8%b"XZYh6ԼStg@'&[7q~+̔W_߷#3"wz!m^[+Pڀr&{ͱ0A|qK#k ַ[ۙ(Vgֳ+[٭2WT.ԱX߶͡SC۴AwT@~ПO!$HWV/,$_xw D~^$%1p΄i;5]RB|=SQz\Pej# dC[lpzD?[V[RB}l:Z&>432'C}&s$}2ì׵y, Ml8j3I ~g%r8zhy (Tz1q*ԐXN:&E+KGƙskmnW%8S)nৢ?5,i4r24z4yjo sմܠg؂}wMF+@Fwp82X6>=Si6gmI gZr9Afx{QF@1o mKO~)&~љZU@eE= X¼Q/"Fb@Lk<;ɇYStwT-c)0N wu෕׋E0Olko~hF11= b~?Mdj3nsMU^mSaX?o6*D2  ƍրp0'dee,:uvM q6V,`3,:26>8{N>ĮgI] ckh3oB!4cF3@cĜE NVFFD׫wLTL~^3%Lϳh+X6?Pp(/d\YR q~3H 0iT ʤ.| )&?l=HXR디8- ǖ8SυK=2֒ax:K 8sU{LˇIi\R FA}wH:l5wt H D/w*nݣUÁ 3aм09=j`F;JR.7oll;ڢEˆ^FcD2Y+냉KM-a$EV 13e}P[* [*-9`O%?]Ԑyv}ĵ-N.چwq"h6_͝QFG 1VT 5 !S1'E =gN ܪC۠A&/S )7$JCx$34"&3[qTlNd03|="],BjGj.q]4#㩒@yuLm~)T`tP=gQJfԆJtRUofI+PR2¤{c ( eu:_H$ \5i9I19^t;IQ6:W`_f^v!nGPykeXj_bRxt]ʒ(fr$ ֈ%ca 0\輶R-<)DyLt 'Ys~Jyk!ESK[A"@ tsqp[R`PV*gEC%DhVLZ"$"MTR# *Dx ASmEFӚ嬶T\81^xT-1 $*EtjR |(j "P|:&/BQWxS (U2b„+0Z=a4 csPDi,j]\-7qR'̈́4|fv9fEdFy'Fp 8WrQH*yDX ' nG-cƏ/Ti'.٠uWlfW6{4%^6"x\)Ru#`xb\I&s"6S x3#s{S;:᫋E+|fnVU(#8 > Ĕ1<rLIKBpjVV<36r$Q %kYݥNN3>?>"OۭmpsUB#ߝ'SƤFQ (HO#ig#W2Bl8ZM%sd8]RSGy:R")9IE$s2JaQiƝVc  B' LxUs-$2zVX c˽:WB# Vy,1 .1آL3 '&ܲAZnߗi"HBDNQ,(B(Ĉzx2Œ$K7n+H; s[FÅʭ--z333.R&nOR-D 3F:7a| t4 hΫX48ݥ0 ߤ̨je$o,A|>WZ7=o`257iK H` ]q+vR+|;iXT _&u}v?|>]~5&7?~_deċӋ& &Z@)hXКԹK0s%p_;'>y@u3ۚ u~|>:qI&Dt%u? L>S;W o˳=1 QC<( HvӅHo5qS3@ Wt ҲϖܩlcD|Ф m ȷV~]1*2KV;GJQ&KC3a0xfAAXbL! DYAvA< Pgqe3yNb;;2Y.cțQGM$49AewJd(pb\Di* #w*-U&T-PGՌ68fي:ƙKN-4m34dX~ ?Ǎ;a @rrx`}Kv l00x{z%m"=KKjT-Q-mU5b;W"G^SEhsyQ-@^tތTFNqq?\ $GMn͗qfI SSDS;C_qIyngs8{$U0n3mwld<;-/+'۵e28&{6)4H!I3d.z%ލ{5iAkUԐ 9DD CMPE3),hHP娀4#D y wBl0LLQxz"7'1 kͬ= Sg^\Wஜ{J!+%hsZ5 Oޙ\k|i=uu`e5tmH`۸>zw?OQG/}WVIi@4 /,<*XƆ/b4v+xE`kjtU^|URRBrF0ǖL L+ 7 3AMȟR!]JӿW 8kVLKH9M(jl@RxWh+1C$*xGNb\ Sg Aznl 9TR$2AB%<#:j>1aJ*Sihl#=5'\v{,Q;g'Dтt>'b^b2"ZsĢĞŀwj.&>#²& 'mobQpBʛ#g+D\S`8gHZe4IypN$"4|ѓdJrd^VC}$&z\Ndhv+'b9Cv-{w$$ 5V37=D 8VBkOt)HE!EUM9>ȒqX.FKx&7@J,x/w0*L-Xr/喎9TN ?".VdX *(P"<WR{134e#ʦ ՎZ[j@Y+zF7V1vMˇVY]zյ`[Π;qI*]?.r~x0_WoQN7iѼ)EK(^1!֎-,EQ#SE+5o8XCϝjLm$!n f!x[HC<[!KCC|LѬ{8PSgw콙}, -}_wJ]_uq,(=&dewc? *ɠi37e|7T8w^Yܬ}n>7xެc1bwpQ{F > d\OCHR gsSQWv G>-X6JP %&2`1x`H>{ʮ:,*ਜ਼9i0:]G:7ׯcbֻ^( *Zw(cܱ}X>w,|ޱp= ʈ4ZPjx&Y,5Gt siZ!\@xI#U>E'EAO | T9׆)p>@&P!N~?ZDIU "wxK Dypd{޾ʜ/dӥϸe~dK^7H4E֤]Ţ|=M.o"i(ÈJd|IsYSy$qhPtH"ڗN5v[>\(I f]^ kRBrP:$Nb+zP!ǂAV R6! `(}Ύێ^{9LM}>s74v8OguW`i Fs/p'!(fJ6`d1t=` @ղoaIHeӂTTz)D%L{AX#{b{C3"U{NM򠀥 $eTG:rE|pZVUUi73`1M;yA~=KO='KhfehiC(*kc0)9*T͒qǍ \hЄD8PYEn99+\qvG9X$GB%i"'"e'aF]XXPkAJ <`L?bO /T{{oxTec,895ĭOFg =B9"'J2F?at˓3Zh9JVy ƌ%'JՊ͆n ]V,ϲrO0λ 1tl_HxӹB2R_"W%7<]N_65j6At{EeݚRpsgN;sY9ꃗףɃW"~햩=Ӗ[P_G [Έۋ NПV+{byOh {T~F!3N/{7SBo^vGLWȕy$WrYf8Ցܱ@GsC\R'GO w^cQchTL1n[?~ͫ׿ɿy﷿r3 _%:C Ϡ+`o=n5.~]+RWߴk]s?/;!C)~ԧrh~V ؇ί__uw3:ɇN&zlt 0 ո?GpG_j hC<X~OsK$?#ݵ&@OXٲ9USPS5}c?[u0C~VD9 $xbD Z3ᕁWvx߿;S: :'%AEN`QB.H_^hk>HSt}>kB4WNyN6 ڪS2*=> IKIx4ZCWM&|E}n¤sͷO=OBj_7|ꨇl<j]pNDYNrQhlSq.Ib5z3{ZђDPŜh $fVP>e@ "ߌD< эpI Y8]n]d g;Wq?nRLw\WqmqÙ^9Ѿ#!_z\>ﺼ㳮l Y̛($ÓUs=z9ټ)dd❉vq@Ᏺ:R3[Ds8HX hȄ*QJʩR|?NsY)A%_eұӷsoga\~'[q;n"S.eyhlYQ7jS{>,kx3l1:8f wǠX?2jیG+8jޱy֋&̳YGCC ?M k>j:1՘{5٣{ܝSSї^ORi>Dfm1sRqpscGZ?jhN%4Y}<|vB0hNW~|;3,?7| V{2%[NC9 σ!R Д҂'\j$愑T6*!ޓ,-ȷ@[Ig}q59Jc'gHm)9Ӟ".d0 #rVHu mm"b [1^'qz5~tD'pNsUQnJ]ߙ89^yoHN;kd4Jc,|V[\.i:EYT䉭Cė1G.~V]5}v=î:_m9{4i4xyMZ B mWԈUVҞ0cC[-Rv| @GU-u-<8"GY JT2ZyLOS76H2%)Ĩ5#9\ ZQSnWp֛S~oy}X%$4%J^ȥvZj',ׁ;D| Q#}Ք:\#u$TE9-uS Z}<-P Gb0:YQ8 *>ꄏh*Y `q.TYgD HOPw#{@|TlJh |V-,ͥMVz=@o!X>Έmͅ=~GOښT-=5a(̇EŖJqQC-E6,@kVpK\ᬃ$] elRM%_<~Qr^651d 0!(o=3 $Lh%%NeIr.RШ,&+Eu\RFPiC4B(zF'"ͧ~5v1ؕWsNJ0ľm7ڤ7kK$qg͸:I%b4Q,\ B Q^k"V~η^4w&r *=[ՖװH:unj꺛ꡜܓ)JXFRJ[ 27LShQ۟Đʐ} [< eC h"M.;-Sˆi)-zCl6y,ҙq)dNW521[S.tdo!7'> -U-6gں[mu @)1XSt otf4I]l5-W\|6.ᨙ6h37[<3,Ϳ9uomu$7  J7.IUjcBk /~.VFB3pecqIA G24 %Vf-ZsYD5Aeg&Nmv)%4;E-6~,Є6e׷+ILk<,@^n9В3QB4eO$. m約ߌ@Z{3`𭑹! p7c?^|,Skgj˔g [3ep;ޖ*!KI8J])NLΥn]ݞGaB~FQ;n}W7+̷wx.s:E~ !qQP}ىx [Eܚt}5WƑw?t~wo~/6g9lJ:qAjIiG4H'Zyɒ&leO0-:Sd+[ѝ0r`H7_͈?K鸁g7O~ uiӳvd֔m˶. CYq 6Lܯ?x7߶Uhޑp#xPY}.1<;B]a]{帘c q|C<O=O[B۽E&xQqc,,I01&*T*鼯,W9} 6f1 #L`~ ftKҜgw/ưJ{;W# q09K>9*qU+V 21t!@쉽&h4FJBNiטH'1 ` I˹Jz$͞#Og 0o1<Ϯ}zÌ͔e._f.퇊t@m+ݎCjdq{!oB3Aת!PE fze:zZ$ 1(Nq9DVGP"*3!Mqn{qpg,"F @ ,KHc1b:ʣ UғB&A$g5zlN-Dȉ,MA`1  fLI0gE|tv䳚~ ج&6B-^vR5.+12 V9aZDTȘ8V" =.wG%5)1X<9\ژ /A FV: Ag'uvҒy;uI=̓#HID4BEJF;M1;VG+c7{ ڦc¯B"M'd?:"2;No{^i3mhJ"pNDK5A4RkMT:e.M>h;Bn#Bs }ѓaH'[fEtv+E)MVg$g1% F (@mPkձG{cMnQKC[tNb nМnCT#"fkr5%cJ5e.mDDCY$$t&GJ6ъce$V%++T`+/By-U[t]Jk;h }گPW]0 c+6 &q$XɼLVkL 0R2&$lAyf91iT).:>UR, ihJ ;tPIR>Y,ʶy[ٱÒO,ـkyFwKY#s`XmΙJAK:j&|VRZgo9$4 Ӆ* n'š@]XCBF'0FA02lK'3ZEC(qDrI U8H OrILX# x4tYw!U4}%lyP-!hAb 1rJNK;? ORM0R&v*FGÙ:V)A:x]|2cg2K|}|O F~ ~;ϪδL֭9 \-`#YsPY\io0N$IzswSy0ƅf\wjS州O5n`co͘,I+81PS>9agF E8sIɔYj~o7j$wKޥR}K ;ڐ~w(Tu&sJY["C?ލ|hg_kZ}|RUW3t09 ;=EC.޿ii%r\8zK⪔e_CY)`M{NjY5c}j~*8{< ^/a/f{+ޮ8Ԇ?Np4^2y$MHRmihY%Ĉ1 G,X4h8[l~ueq$!G}"Fm\k娓r5-x |> ɯ+s~{ٙGr-/D?燋./?^x~|GZ珴ӺY,Ƚ& x4?І54Zah mz[g:60O]ay'nHoƿ'T[AO :kMtq=Mzٝ~Qոg%H#/gc-B0ˀ@_{N(?H4j#=D+( F98+N" ~0 h )ғS  Q`,Y?vh46ِ%2ǢI^hICT> R_QƠйR rgԽwd!O2M@S@e} QKr!kTY푐Z BTrk4+$s]6`.8Vmиle~RknC4l2W!VR2Uc֕RIT]}:3d(Wf@ϫ^ִdN-7$4̝B.#S.r2ZYAR̂tkD}< !~mD7blZ+rkTbZ{1~n dڸQ9vHȂ.\v7/AhHKZG0B+dv**We5WG]wSOT|5.T7їo l/R``2er:1+$jpC8FC2uUnvo?^rЧnڝ1NfIP}z*Nsm.dUpldce 99n Z8bX+T6ʘ7Nhb{oX:2.}3$e0G thu8*][+>5s)&}>ҡay⇆ZU=:^^jQf!lxW9$R, HkЪ=lfuQ;E91%Dx VQ׏Qom:;g݌pwG7~8)ь, 79F_r70|+10r]NOCSNSX$#1AV(nRRJF{,uʜ!Fo u=!1L9مF#r3]b>E&;-h-8oU 9թy"%Te\@G-IPXKp5t3->I/VꪍƕF9„DQI|均eI&xɏ2I[C%_<;ND/OQ6VRȈwAXG(+PR\g*=G_^2~_ރц 6yp )uroX pz/ |)=Wg1czM%}v]7}OaÁ2MGt ydqjl]t5!#ft-Y"WNm!/7r7 K9xmRYIidr\9C%bJ=ڢtR:wـ ڧ QtJhZp#ХIE%,H%@+.3ύJd' r8rv͑+ob]ianf'UI)AlɂWٻFncWX%H*?8UN*v*/Ub"eZ:4Q$P$X{g`0_7Xd$4FKxfe4LXY.piD~>V lŵ—ARVXJG@j)HJ18 CPd W0sD +UMF#ozI%B#5UAv h'$ %<.f.aE  J[鉖^Vb ڇHtd ĸ^G~"܅#gw W }JWpKF^&y`5s7g~](3|Ty$V|^Ӥ+V"祳mb4'e.HC{%Z+5ӟכ=:Fo3THY"!h>y&g F:SyͦKy 6ّ@47?mn"j WGū؉}t}qy"lnor?ƛ&vgUfsS%޿mgsJ֡MFsl[sar>eMƻ}|w3{{8ɎF8Ȼf \Zov>iݍb{rvyǺ[\ݞ@hK\м ϫiq5L_{-K\<}s2;X-EoG?zq3oíg?nस>mKYb% a 65)2.I6 j]ķ{)o<#8?Z?h mɇGD3^כ/ }<:ɵZj:Z(B|N͞.o?*cɢKF3-H^fe-zS$`Q9 m{Uu߹q#8mnU-cw@{D9^5m2^^jͬYxYW Ы#+VU7V\bhea}W MA=ˤ}{ YږXNoz^:*wJZt/cq8njn㎅gtr1@ƹXAy|waׁt![O*x!i RUbۃl s欔LLܿl~OܢgNYb J^V5͘ >'ʙIq8 ~zyx2l؟~xa>M ICC}TW&5XT+4,Iw @Ɣ G"E^E% IGmm {:VuJjS[]Hw}iCcSvgzS4X$r<T48cАxԈHURg*B$IT{CerQPa!-X\.gZ3&*yb kp8q>3uu=\;[Lns(73ۯ# ak a4䝫nQyr;!ʾ32&ɣ!"3\C$\ E?1()~4U FJDΣq!0YٴnZ9חņ}BDܳ[ٌp1N(wJ+S3s_Ŗ60r^nn(eC8wN NyRrT9+)`F.A2>_dUsic\]C?zTGBrd *xciJ>Hϣ5 3Kq (ȑYN$a6*<1j.rb /gņzwf [5QFSm.J g '` o!R GCī TNI+PNYPerQ&1yc 9Pq4!(``wBbPOO~RyO;9xT<3C1 vJfҺpY8N}1xa"j$D"WAx/OM%YU60A A(OLtA+5"­fdDv7M>I_dTDa[l쎤T4)/?9+6Ǡ> zU=ʩTbR(\ȯ1z7ɆAw"Ҵ,pxvi'% $Tq&l)-N6Y21,/Xs]"\54`m728P^[poEڳ0YYl fq"HNj P&2 KG E hb11P+I6F \hL10^8 24 G 4:/5=-r~Jfn菮.;^ˣgK>0MڐdRJ$Ϥ}!Q>%FQhrz!$Wd?hnK- 2=q<*2%O^h`Zx4PaE#7E!*.hvFz{>{yD;GAUGʺS'vGO&zy6[qaJ "do]fGr^[\io%4Faq%1c` [㟾1I߳?4䲸ƞ_hY8{kdMдtrp #[rq ;7Lq̯-E(Y|lDC.hn OmBS8S>.g科 1lowkۅd>K8.~׮D|s9\ɇUƴT9[uvhӣ@;׹j%G菽G47=6\ٕ M|XgOc\{r0+"Op98V'ugOBn馫 0ͪpԮQ8y% <ٿO<2CvYNrVm\HXu4 _#F'XnԘ-*g{{_~}׏D|~;:FOPaCǫ +֤Tߢkb + / O]Q֡mm󏽿0?nbɛNV~]y^X  vT]G !^ݸlGD>c4|e1[S(%pnvc VF<12.6O\4Gnm'Ē#x0ШX.0p~!k>HSё|iBKi^||nACj\˩L2qp4NL+ Udd_uܪ, ÷br^ox)z4\k׍o, c!lru㔐dԜZ5v%W_;hMW?J2O y@4N?{WFJA/mc-{=eуYU$KE)MۖX8Ȍ#6[bM7\_`.1#.1T)b+t@HSV?9SX;^s%cmcSٌcʩH8]._ |V\(S ҍ\ .QJSFё΋IVX6AIr&8uLD6r:%?T|~J*#yDȩE piIHSĀw#ms C9&r6ȹMVIGEdQ0A[2HR+kHF g3: Џ 2d8#؞cCbS#d."8AL .sqW/<_yCW 'GnyS#֥8dS ${W ))78ZNMYT <@>RV1f‚l7 #p Jmc}J!,+xnReZf8o;m^,?U oR #XF'i`B9jF#C4YN-8VN ^] 'IyDiU {ɱ 1q lr#5x]!bK|H ƥ0K"2! `Bh 5(h >AxXAxAxhAx dAqiHbytpuPFK],eA8MрPHY9fBAaSZ;|9J o@ܝ#[nO灠XȃA;rOT3*gRbrSܟ@&D B+-@(%SJC )N{d KH$,=AEJ.DpCLDK9 Q 5 f:qd!fn>/GT햡Ͻ+L3rl_/b̈|7d6o2^TV4 +g !9B^wOۈV;=Tͦ{]QCdI˫b\)ZoO7qn4nt@Y Oucčky$"kY"M'?g?guᚢPU(МO9?k{W9NL-|wfێ/ Y*Be_nB/CdHʲ"z;j˵yƾ,^VMLZ¦it3;2Tܽ\ PfFRqNTSHpz&w^k@U?_a}.$ $`R(}vim.hy 594 X^m"+y}&n.s{Vkli㧝Unӊyʚ L,<2sps Cm1/.n4k9@t9lZJR:^nXRm3tnՀ_9+74v[Ƣc\6#(Vaί:߾z]n6#HqϷ=}̭&>]uw'mP|؄٨l){Pr1?Op{9V0#)`6`|ӞdrVlvzX $k8S`u.{i\(dGZnsgB Yߦfgxv~62,^%lhp zCgzıf"8|f|+݉IӉ-+^dw8?Cwg o8xoT}a~|T% EּKUzWdɏ+ZZ} xlGƉM@i;5g"Ejy?{2> -hߖ%9elU2EV3 a~0qjaa 0 UXc&6a+~|楢Qwt*۸ G2 TKF3|[&$a4JmC(m8 )kH \}N;asM.n&UL$I//_:Kl(+YKPZB>ֿADCCS dw5@( 2B5d`7:6Dil{4cB/bcziRf5;yi+[hJq͚a}{wjPyEC+^fP8Byl|vrmeXQe<{v(Qx*!mRI{h獨sOۤ^eˆ]Z@@%Ą0Nr+=jܝm_D\ZC=8ra7NuBBze-hΜ0 p^ٰ?.`| wxS*w Z TQٷw&1-0uYVCv|?2=o_`ș^*|uPʼnSU, G1UJO\3]pTF&X*xhs6 *z$( 9iWk{QȹߋfUhiV.uި _Wvb@t7NtnVʌo@P]T,:}+}4YYQ <8AVv%,V{EyZ&ys2]孰*-la$rx &9{ k>gO_^ds ]uW7U~Q"}c3dbV\3,g Gm'Ófl WLzR0w $i0 "'8mpk 9>$ԜIH`X( H& gYA <Qż PN$%Ӓ+A:72- S)K^ Qp9(P,AHE͞$FiIHD#Z7x"WQ#vHe1eH&^_X2SΌKAŤ `s+1昪t ir!F~vRI_1ǭBa`4=7,@rx3 $qQ3ƗwL'ZQ>r6N[p^뼴7I%)1-|Xe6$Oz[jtΘ׹28e[}l%>>(똷*POE #!}eAX1pfcPHHAF-a C74Wa,r6wI}bYf{$FĔ1Q1hFAJDig#W2Bl8ZNR?SYu;nHitL+&+ J3)<I@n# XV{=)vN) 6-ҦFqp%4 bRP(ͭMM+Nm>݂Rƞ@rS~̩jSzخGUn:͸k{Ӿ)iK"`ޛq3+{b|yOUݐ(BVvSg >Wg%:&lw5 W:^W5ҴQJe&2:>QHR&sYuK^R<|YW>\{pӻ^|e~˷W_.޾zV'_1]y55&B׿^پktMۢkMuEV3Ĥ)b5ۦ J?don>QYoE֪,kf5k7 Mg~6ywW(X\uCV 3+2-GX}>>1[Sy? :K2dLg'P-jؓ, *0wHsIMܬM:Fl i,1EA^^; [OS(xU32ƙ|5Paa\~|tt܍r0GtU[hOБJ7_sk XT $}M{0'?WXv1lBQ.s>7eӯD}=㕦Z][o[9+Fv0s:ayrr5x$V9 +1 :B)Pk;XM.RPм.-E)?W{#k)#Bg^!TB&CAJKv&j5Zfn'*IE$LxͫadXx2)ǬZ eWMIT7BR ^#Rtぺnƹ>[Wh9,59Ked۷[7ׁ_dywÊ];ȓu\2lԢ0c[T5M1U_!d :ţx܂!]^}?hB*t eR7Gb#n2L &o95dЃMl$ϋ5\f4[xM nPz0Q KBީ avbvfvcvJc XdC2BKJ[C&C(B(J4 )!I1D!`J[+֌t% +,l3&%vz@H#5ZNhR\sZ n|>Lxe^l'*:<dȜ7}}6A{/ګ?- 5H@8Lz@P+zz㦷P.\H^9Tp$VEǭK)ސv geu| C^תV+g$Pz֔hdLAPA/O ͓Z/>5=Mrv?4Oؗ'!3Rau],8J9XbqQ\oRRʆ[7 oᝣ7f4S?&;ܴ]UsSvHbfx>)H[fcz~b빜1&j"S+@hevs4-4GmP G/jj;|-t@[jH6X$M> V8Ww(*<ս0Sh8zO: K:@3«=3 de=GGbNe0rj v0ƫZvCDʄA:R0s9Ts=/(ka*j@ŕ@>)R*lL:2A:T4BF h6}@dRP.@Xx:2XsJN#j8⢨"}_RDoqOHCFr;!8Z.:&R*ȆD%lSprBlUK͆tPzKJF7[=fJ`lr!R9 .Z ph=`@1J8@|I`x*9H1k_rO&ǩ)ɘ))64j;>1{~t_;yV70vhq9Fw Kw\#]0`O0W,. üX&kƴaI{vVkPR[ =@nXl۶qUW\OkSr+.mШlˎ0&'7s,լgtr1@|ĽeDd޹xmpu}u.by)/jE@Ά ڽaθMOx=֨ldEG/Zl/v*-mu-IG_OE_˹f)oxuʛAQ~p8m8iKvF" AFdyºCg|-JR̒Q'Al[I[W`"Otghq}4h a۰ۃW,s0)QK}2g \Ȅ,F@?h@JO\ן^b7wBCI$v3`.2!S Jk4yJa9dDAH#1S:4ŮXTK#PJF#5C^'YVsᴕHv bA\yNj ;흉c̏)qF&NjV'Ѕl*$GtN[DJ$LJD#|P"8B*@Vr*T Z\ `C{q\^r~=]xu˽^Y;~Iu%Cۅ=>xUlwIgw3"!EL2ZP,=Y@rN(UJSS΄(cdR) X)ʱd 2їQJk"j g҂[QQtaT]BuSr| =W''f4<C..&7o@cPtʖjŬFTZUOΠ3٫=%& 2(C l &j!ad JidJx\Drf%򅈙l-cl!aܗ]ƺ! p0.hy7 a7QՒ^TÀZ؆.i]ҺR^@(Rl_q+#m0 m8Ҡcbu:/s0_w:(s&?}g_rv}y3ky .AAxUкw^ynA݅͟>2V|\^^Ӈ򵾠Ԗ{Չ}w}*z\jok|ǪPFɊ{K,j/,\&~d*/}XGtX|ϚB/^J%SB*r&mb&g@<љȔEc)T.)s/%z*zT rk._/ҳkI{oDg wJg wpg wpg wk3#mG -Λ;[;[;["sllulÝ]vvig5 y8-ڭVSo}v.wgv %s;s; +#jҘH~݂ 8VeKj0qB!\ xCF}:>a|QOѡO2͡cޙzښ|Tɧl y3WXHR%_^A|'#K.Zgu X‹ RzMt^KF`2Nfu3=mW@]A#{rA< g ui=e7P[y<5_Ӧo:')u3G8*7M}+Rx{P B0Z PY!/{}O騙 F{K>seH2)%Kj_tԄƻ$ttr%l49M` s?|f,-Xhd' !1<(A)J^%)礏j4G(=gCrv$c3;k׻;Op>=S5fߵ$rY|l1DF\\pM:ˋw?Ly"ֈdK%{}MOӶhۛ54ߡi%{ vYLd״vԧI:[kbcכ/6wk'_:hw3u?tWPlTyYZ-KpGKCh΅ʷqOʥmhi1[S %ԕpv8o7w{<.xKQG Lxe/kGv|%gJۨk|XT$ %I C m5}:2[OMhN:mr2+I22$>0E,҃q0J'HK;K"P'W.6'synR3jlV[Qix |JhWWB\NRS[9OZWQe7I.Q1 mF}.0П{NC{ٻgKߴqOO Uyh Y0ZAc_"l3Z%xnU7Y˴6DB-MEɴX'-*D)iW:M)q?wE 2۹ޕsatM,I-p؋\Z׭Zn!aͱEGGBG`Ki w:F$-}l+?Dxaq[J+r쁊&0 £6a뉄'R 'r 2Ғ/ML9HκL5Bs$:y8URλ88L{cbT}2ضL!Z"ՁHOȱ8;p4rwJqEܹχFݮ;@z|jZå9(TvŠtbS#Z2S-IHj/PhFFC4RKo5ɡAJ/EJv.&v ^= tկ:SWֲg!1kA:M%( $ BY)D1Nzchs)㸡>PBZA=#KA ò/@8;dnhMANԟr JWT%ʒe¸ !{OXn GsiOHɷĤ!H)')ZCuFI%s&?[7 5D}VW eDHPX'3ԁӠ7Nq1 Z%k,*(_=֫S}:FFMٔ |V-,[Vz=@o)(NdHGеW}vښoWǁG:D\v|>,^_\ȾEL(+КUW9" I-d<V{+ƽG-Цe^R˘|$0g "4){XD)I L]2$9,4r^$ gsT%(|PLpjy^vicǫ1d{ d1ՁkrGۮ[s$4aŠσrD$8UH]hYJR -jhe(d^'96)HCd*ts-J,Yhvfv0A&ԪMs*9Bɹ簡Uw&X.Xs:t6r._{w඿&@T}UHɽbH=bH{|0+U1Ru O{@jVB@*FRR>BB2/*+*ze'q H /zA0b/$N"H&wl O*m2|DH@ö2[nnBWVS:ۃm/7W Cf]^i=G 11)cl"e#[/yriQ4שr"PpAJ} /^mdw!쇩CR)rljT4$A '"0rG:fKQh18N :g2B"(J4r b3]^>\NV۷n; 쭭}| Q|quV7OuUTיApnمwz0u⨭Bnm̿dA,3\N/3oe/ox4KDz#7_]KFM^mvm6򇟺uY٩ԮE{'e־WYXZi$|0KVA|q~˟Lj\j/oKƛ!߭fV1[•2 "y-\Y€'ANBdd l21AKdRxfM$wUΎz՟_-Ni7j!9H6 $—%dSw9YmDS*!zu_IMhJr B'AY^{A:!,@Kv` :U ji:;i}ǂD=̃'hy( 4 Ry-U&O}]!v3MhJ"pG@cQY#vh)sA@+͵'"Էqj(_-yʷOŊF!J$,晴)j5|(gjJUc ƖǠ$QٯjR6HY;~uZHCBlB&՝MV+OX%񤲉ؠ2z w4NEtw((l)nYW#I29Y( u`Jɬ0.$" Āg$RsAp`QHCW\ \護(J:&PІPϢcQF絺#RΏY~5N6Gy!BJXv9gVk'i_,L .`V6爄g4t 5~K&&>x!. %rƂdc2['Up֐1J<)BB 0Ro8=V[*4iXzt1[*.QE/d˃1 Aȍ%(Q6$XJk*$*9_o^߿z=^+uJ30M65]p ¿{tAxMVijoѴr|7+NPskaԧ5h/romHwx5?\LB!ݸ^\?fל%J:c/.~Mj>hrU^ ڑ .!0Ӏs761WF.xW ߃>f+p㬼Ҳ.n`hNucd@zl"SY+T0]{X?o6]2ϢI`QICTR_AYO!*ȝQwӑ|aBK7WN$ߢ Ǽ&( | :2Q[Uh ɣ%TKW&_S>{ߪ{.^V+0`o2VVsYgRe]h: Xl27[HTS¬RIĺ+\v¹$_4#{Z.ҍ{",MF3#ӑ)UVv)ufAzS7 :q^Xy`u%=@tŭF rC(ţ8ZwvȲuoA:"ƝnX 8ս[MДiQ.CĮc^9ǭ>0'9I>> FCwUYń7FfMhxf*=>z^ H}D1F(8LYd9[f&6mH*r|/ܝG?&GbI 1Q,Vdf. Wy]~G}Cnd@mv@ j#u9J[T"EcH;XTjYVLsCU LFw:9εs{Z2MNoHYlN*V;ZwkJx4uWg~=|ʾƦh vEgYe)9H+'WxJVG- P5RhPޙ7^1"G>CGQ?d P`Z$nME>xBGXE:xp!+K[,ނ7>9I9tIBNSHSq FYEI!H?'Ez28;'sVW .V1ʔc}(hh#)2 dﳃ=NU_=d/Ws!EJ9<mCQ-# dEx0tɒ,R2+yBI 6Lډѫug; }Wkq̈́$r+wB}~`O ό/nz A)x>s祜8wccL^4(730]r*v @z$) l@#dU5NC1֚ .DVX˄下PF*Zy!7ȕL1^rjjhM'ɁGÏw§qLK펾g+,Ї?,<84?:^Zt}9MA;MJ42Dis*r硃`6xlRH)pspIdg]R1.HLIrU&%V,hK)csȌzc2/i}nBmtexY;;/ip8_#kB ^ }Ryy1⭭ | 2$or=AS:5ұ6.{գi[')?۹6l.}B.~ ` TKfVzs4ɖJe6[+ *OjUr<-n.2?":4|Ks{c)soQk7eiRYgq:,6\1,n>oi}GR(MN )kPkZ Ik.1Z/XW~GggeA[ުك/]moG+" 8mllnw?-7-jI*w!)JֈԴD0`JLMuSOuTϗ<0Zzu~`,<|nW56ВƵ Eyw_& [O=?߿_}ku-,;oMȈ~ry5:KU ݘ?!bnqj,>,KKhtq`A_U^Q,sƵa~ҵܹrhuEZUKW5᜴ܪ*eћɪ2)}ڳn}{N[-ɉ}=}m7_xٞ4rd,jqZq868i07;×N'?>U"*ОS[6{l4ׅ/Yp/v6iZ[-0dyyyȗaVp_6cuYY؇@oFM^]\#SE^Jퟨ^~OPv6/oRw/Q6R"oƞR@%Eԩ+{!mxSSݝ=v6S-c:Y`{bZOĴpiai^,bzvr>v:(U~6b6`]d4q檌Ud~t"Q ĺ*a=jH(#K(uL1*g|R)çKZ-wVtZmЬ᷍cx^_1k?zR6#ŸQ=-I4 3zXT~WB\q\U͍qʅK,ݩ.59+5Zq"ڲx ø';;l սp?1x>"DS1.9mt40VGVjE,UĿĂBZca*F0u*Y1RgNQ|̺85?L"UX9Su*3r7>fvr5+ͻf{GG)*v>rEY5*"O&KLJ<[|}0`"7EFlIDq.'G JQds2H\+*Lx7Ky,lfB 慡e(zܳ/".a'i:6,_;bǨuI`YssM"+S:.\K'ko%dB hT0*Y-HE(7%¢uwk%n\>neEm O Tm,4WH^AڸBeY4"p:PHE'fKb棞:Ǯw1ceJH!T6p+iPV*">6|s򴓧<퐞vvhQ,ٮ:wլXIKndbo' WD.Ey!q4%|L#.Bf>'<(>F;1egqϳOՅ-rgG7!u""A.3x-XNF6;?lS~pfd[rm}wx :z]ؾ4|SrSw77NOmrbOFn(yvp7qR)4:_/;k8uch^?.k8٨+}Wusei7 zx+r?>u k#whk0 n?M<>%Kۃ_gsĵ=p Fov-9/r1Z\5t!Z.\c35{5k~#QW\NUߒAݯg0!^^瞢u0>ym4;χo7DWV:}k~s>!M_+7Qg]>/Sq@c7u`ԁ}S.XQ d)j5h8eҪ2Ӓi52g4?./T) ᏽ٨-)ꗗp `ٽ~uM?ذ-F\j{oj)q:&na:GAP5|^w,[rIݫNjG֋bZ A&}ꜳo_/S'Z|Y(Mp!PHR JD' ^BA.7Tʹ'jQy?V{UWwCKJ郞Z}|<9OZ\EF1ϋI^eW(SOZd/͉i)wڈz] ֋@Aq%K>XrR2J MhB1IQ$NL\˒aR6B!5g+6UZU S-*-TRאd:Ӑ-uLAHcf3(:Cde*3W+'aZN)aKDHFÙ1EaEG(%uJm"²%"fl{;,ØWZ4rxd, Z bnX4JHXH&H <q.ČWiIgKŒltd~\./ / R"!i=RK--e&D@q.KRYK Q삔:  H%5(,.(cNd* 3YB: is$3` J2P(f9E DGk̑[u n{ LV\0߂j吰N+4`$|x`aZS!MngdfN 噴M*R!' [&TU9 EBAXc"6|0M C; ZzwL[6ۢmElKa ܂NǢd2X5WLEHw@6Hr݃g H\|Ka[t\:H'2z}Rs <)fP57].ǐ?g0c)H`ĐVH@Lh,MQ(`ybyk@AGLS; r `P* akC&;ʩd,!KAdj\ʀA`I p45Y #VEPgЌQ".ʈq")XT S*jv+ !XNDugJ0Ҧ:s4,F%4 3T7/ T u- Y_cUXPK019d,}יt(cuaHt mZ Z4f Bd?/edv7kjfBs$ D][颅$Áf[N)`wMS\X aj(Z+9E8BV! m'Bn2uQΦ/GC{7oJ3Z? ,h+:%\2"RB gzIW~VG`ڋ32`m6%>wd/ERRQ$F"ɨ̌Ȉ"% X=Y.S.\EXQ_^ U:* OڍºaL̼s0 qreIo`uךTcW2eD0q5k"A),(i(!e#)9[| \4(kuC`BvuUDP06c̬o"HMXE:.E"* z lQe}"`@x',BpjJW n^5qsqd2(P_L.yGB֚DgtQ#h߯n|>q2D EΐaAtQXPMnv-Zڣk{Dq@ TgCI ($" $H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! I Q ;'gtL 8yIHAPm8@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H]LC6NR` @ .cgC)C-@8BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $B@Q'Ή@ gC< Tf$[$F# $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $! ;~TzMz}L_$;fm Nng@ +u.K'.7.zFj7-q4J>XW}T&! IwrPAT ¯|a0KSKgap>`,#JpJcOm 94kjJWqJWi:u+] ҎW~571˴̏S|k"c8jF&cpޭcer Q0.ū?hLg݌4/t30(/ovi'*HYT\O ܘ*A\()q6a04r-bt:IX?!'JH4qbJ"ׁ?կˎws/?K3lJeVd@'4:  `d"cKK>.J=u? =KgA!*"t& ep#*KE|R\RM{N ʩ9s%"/*:!b jQ3 LÄYBn`jZ]hn{򍭀v|gM-QE"S;.4]w7߯Sni]yO'N4@pb]$hIPJtf4ztE0*}~$5 /fƬN5>O1G?r_"Ž/z$Փ b~VUnXJ +[W:\ԔkcBf^T}PbrRǁc+,8XD)RM9xyO֫fbut9ER fs5bU-#t{ms_iύ79 ӿ=o)_ gC=mZ*iE4T}t.[y4 7kESV{nQǭNݵB'3q48̼ops2Ճ9DXf/6^cS%JEЛy>M5Yա_ \g32Џ vAwyxDHQ`z_%_#墤,K%2{AhFe:AqK9A bN׮O:+c_ LXS)C*kN MNi<:%穛/8S;Ř|78 /?T_c[IY "uU I%jrL1sRr$T^wώݝNj4{OKvl1xۄ6[[o8j6'pώs(1yNӝb=zXp~57Kb:|w;8]Ou:_87P<-j'㌥}.塀<<_Ӽ$]Ci>p*hr7XG1!r}W09 n:sٻp_ǟƓªPNy Z➏Ճl:]7X^h>R:G$~56ÕXFo9R[RtRiA1U`-ΆYw.2H+*K8zvSƨjK21ejRGu> q'ȁyN=% MnYSz^w4NIן+hnjjt0q"bΥswvi ॳ  m!A%CRK%K2*6nS)cJARmX*QIUJ.qpLR.e6%1ս49*R,s9 )\MgP&@Z TٓJlm+Ƞ{><3CcЖw7d[jΚi;!Ȧ/M/V\ϛ6:tBnm6rhk_MV?{R}n&pusj^ylGkG+aޛMp%+bݍ/hya:liovw}ɸ7uۏh8y2Vv]oWSsyK;ܾ%j֡蛑GͥeOz'_ڼJmռŽR$/d.Z`Yk&1H]GVuZ}FGr=|ިٽ@Xt~`VZ*KO1ڀ3^w;[Yf;.DC_^-4b1iW̓ qUa6l I GM]QA*|{ajemՂ+vovKTw_2(Kef>S;5jh^]YP)^X)j~: Sǔwc٩K.6nնWuzq&EZSrw=-/K ime@, ],w$ܾܰr\O_*𚹸"-ז)WW9"+eNEf`]Gyt&=wu6*D z_O6εyu n/jO=m7p(/#WnlJ;t D-J.9U|bQE, l`X϶zu|\?&4{U{q}T{cjs%uNV7^mO)$טjT{-FᲗ 2ɷzɻAK %B@w8 /)r)1}p">uv/,ww>-c4Y5=#yZ`(; ;˂[DʟĪ eR9r>.U3,DFHPQrA-z]?$lBR ~e\)N”,"Y(J8g( }~|3YݴQQ w}k~_>vY'?nm+Ï<fd/Gy.DaR`Ɠ2+1ӌ(V#qz&B0_-38k= LG$KoV,s}r~Zߎ?tlL:ʷDmoVQd%;Ѡ3&X^rM4(Eq1b_=,f@o 9L1ƞzgi3` aܮU} u͓sOx_j9#3\і- RF}dIgaN` t\ZSnaܫe֔J>e֦*mʤO>'[&'ʹ) -co춌۽fSmjgbW4>\nܳh,Go/ ~4| gncT$4"&b.&cΉ"hh?{Ǎl_vؖh:` bl iZQf$;1xF1K{lVTXǣ$у.u7eB* MMk h2{ mi:Y0X8[l?rovѱ6iV!;}6&cbC6*Yd6N3RA$<𨒪j/!2䒼 ![ #Mlp>fY :*jo{~!ƽkm-"V,b5EΘ6֤6kE4r1>2餹 C 'pJoŶgo4m5Yk(EEgT'jyO*ZH82SGw\?Mb1$a]9us9i]gDҨTK{1|M/Sq.;f#@\?1yL6-$z(,=Z Sk3#!Xa0r|6!8!sT2"HG>ecD@4u&h4FJBNiטH'1 ` I˹JzBnFpIߡa5q6H%kPqrv`e45ka&'A9;|f|1 iM#ӻ4C5aT&ARo/oQ7di+ Z&IA1gϵwcD!Z/: P m4gMYDXJ{G4#H0*6XWJ/ )DrVǪzXv€m8ʲG0,`22A&2)<Ì)> 欨gPzfw BTFpf62Y" ,XJg DS*aƿa vǿД xD!Ƭ]V bhUAjP+4?IK,HtHyp"0tMPR.[@{ŝTb7Qp/I1ej$F~;AFԡ)=8Mj zN :bOZnkFDHI ٷ[ml|-r7P`"4:HR$:\,Jh@PbjJ^c GvTԗ5tI_d]f>w6̖c~ܓrܓHEBBK6ydRy[%+MY)%N& g;)yqKoU]K@&߃&0ص\c(;מgg_D"+j2MQJ<(,' Ó8 _ț+[JD]:?|/4LY<]x_.*0=s!eL}rYcO/}{C<6 9%5-d0#9%8Ec3vh!n:MA:,9I#+K40K2LkE#O!p-;*(΋_n0JHipzC2%W-˗-燺5҇'ӃU=E,m~⸔7]B~H3ͯ./aF4;?UpvSʨ=u Gdxַ^J~Wc7ZvIfmKRf`m31ѸQԃL>|Zt# U`wxu}2NٴБҰ$[_L~A>vްcqSMc⟳ GӃ0<;{{|S~_oq{}W4zBbՆөץ4#/ 7!Hb; ..MY\~%N* | 1[S8g7-p~~f[sS  Q`,em~ߍg#'/۩fgC&z- ' Q, K1FBK!*ͩ{>6e~+g+~^@e'( s \K+t)FU {m"^+W|m}nյ.YV`o$Vq#3X'gC.0f zV}X')Pb X)˪.搅?rx@n Rs|Qa˖2w!vR2ՕRIT^-pNoںMC;+ɶz'-7$%;=bNq!ȔrP~: ˽shM;OF-YXE 2d* 9UΣ[K ^3H{Q#c}Mײ9w"4cuv楛9ďZٵۍaWeorl9s _ԅcI׎GG݈eP1at&JgFUY͕o~o p=lZGku;" <m\3,Q*IZF&Czɍ6*r ^D Heʂekr4֙$$"qLef+#jl5.ti[i>< ?̅ޏ Bبoރ<X-vlIcH`Ir+y,mJ PU*NsmLdUple 99n Z8RXĹ_oY=y]FܳկGE{n)ftY>g) ----){{{{{{{,, נvnۍ{qo7)׸6g{qo7ƽݸvnOjۍ{qo7ƽݴqo7ƽݸvn-ddqo7ƽݸvnۍqo7zƽݸoۍ{qo7ƽݸvnۍ{qo7ƽݸvnN]nۍ{qo7ƽϷRYVTK9haFUVR$;d;I\9C%hl5A|tR:wteb#Y$$ˆh$n$!LkRfW!-AxE*Z9p4Ms㹒)ITIĹ&M+KgZ#j䇏cEաPe=/ɸSLd:B3,e# 3*Pa]e^%YfTrP{F&rUjܿJ@bi`m%iW{?v= t b>%_aV]n7<\sR+r)]V^uRy-I2Vx8 e)EcL:lxW9$R, s4*C^IfJ`pAd)z 3O*cMy8*EjqmBa-#<+7B^H4h71(MtikQ h!ecPg<9/S3'<N|L<"leV]U0"R 5 "=8 ^"跸}ER}GRER2)Uӊ`7Trp\^s!7) M1S)tL[iʁ1]n1G o yyρYY|]oO|skD&< 0fTr1:#G;9P 1rac19ITbTRrI5RMJ>%r) "H&  Ze6ZK :|Il~?R]kk,pͮz o`8il-)7$7%r%iA"Vw]zfǟ `v n_Zitގtf}t- $պyYϏw> ޠ煖h0Xj7Yۼ=1bx5GnlF{269g-H&,٧\v{TI/jnn~ XT 4yGXk$K颤> R(J@kI5̘NS>搸Wg(x+'e o2 ~o!du(g4[os}|IV$&8RHpADqk LY0 1,qBrw{ɓ"1z{u 6ēT ^*}eD l{#yհ|:-ڇd/>ݿ_lz*R5,Ȳo:9JYWM;y vKOyLjfڱ{kX [7STY'nǗwgv1fqh _:3S1<Ѭ卩eIo{vVWma5'p~9v2.磓r$垘;I_s zUeጎwLDK̶#52˭+x0]{Iy]Z&h# |&:6qsWxLfe{ ]zԅJ u!:׶k/|<P w'cJ w|2l +<  Tiq&hCrѩ#ˌ/BfNXeu=T,+0dn`VR*77k]ݞ9 B6-ugFFՅMG/\V;?s@) @(ƃ~kȴ! A*G.7}Y_$$&3: \q`׺N_!cOZl BRE;(cBZJGV9kD]m7k]q_lg؂PoYN.&2}tX Φm;|45*03q:rݩ/q 4}9< 1˾O3 Ɖ̥6D'-$C]Zu6̬S(nx~l[\ ?YD oPU)iFNY^2|IQ #ci XVi4.VP"X IȀ"0Z 5 !S〉'gIR8uF.$+y`!I`!idTc OΪEYxr :9S{}S]S*Ʀ,hXJE]9{}Fȵ穪d3-8ژ(zR*}6gMiSgR%c5r6DV[B]YNyrW.Z077T9I>_{!(beI1Es" @<9FD*zT&(^G>r*El me4EQݣe.]N2 .)RqA<]kul+Mem:!ؕ:Sc\6iE%kʸT<"cUIUYi$jȈ D-:dPD F2 /xYEPN9aϽ`<V"v5Rm$II\ֆh$E@`eI3# \^j&Cqao9l;iS}i=4)*֪9ʰʟs+<[!Sbnosw7QB4h(C)+OA 6pg/n..'gfxq E!!@ བྷɬ_J{۩ 3i4wS{^(}0[ߏizF_a/j~7Oݛߍ|Q/Ba'w}Wn@p486<_=n~S y=Fwe85|- H9#^:O\1sUmܑo;rF{tDtGQeȹFs$Ȼ&$J%)Jǘs-s.dU˞l9Xb .0 P;KzXœ4E-red5r6Ic*nr3 YOo>% 4uu Ǽ2hu2ϲd$85Q C::=X,w_3_h#_l;ueUBPD|!؈/Zy*Zg };'$ {Z⪐ĩ+RuPɺs) /@EcPe`I ^^.INGp¸Qu_o7LM#&F1u؅es?Tj'f/gwW'WϬysQ/I ef!8S F MpVl^+v9.C">zɥ7\0ch<+vt虰a2$F (|ꍿ]3ϖc3mQuΥz45%}Uy÷? +Cu7S !so/Ѭom,K@;^%4T ?lV>楯 K)N :S'!LcG<Ō9m#f^7]/u[+^GJzqJ*;*q%jڶWۼzWHdU!WSW].JC]'>ZGL87ס!(??^,=h$~b`&+%BYyTʊ(2ͬx3 5aԇ$3ܙ }-I%lKIgl  efSz_§o7%)S&I̟.3ɉ "$&L&eEST22wE>Reb2$ՙriRA-SUpU}ۿF _[ f-9;Oŷx*.$T*hf)#PµIJJ|6\(T̙:5i[SEkMWJG"JĶ`roys{%73oBE" DOydWd*:ŴOscTsbtC옷 \6MS%bvk=Po s^,뜃_ߠΩ°ƃQHD)-2H2 T@2O 3kWq黜$"`UHs\zY=HYl:rJuHoKȖ~qO_S#=a$##~rPȨ& `Q&u:&$ :$L(0L9XpHGm Ԩ`"^QS&DM @"uPL[&C)ƅP>S,6{08Kb(qR9km7Pkݪ,Ic2 p%gи=wҘ}8 o J{HP1m2 a &V9Zb6U|N$+P$A %Q ~Qygb @ !lmI'M@ʡ>T1 9q2-8.z(BVK'X(_1%kb,޴:ack\N[c_}5EQlMtM+f\,"c>'X C[&|7B %!*a*e!`YuR pFx%8 3GW @WիBo]Ze> m!_˯> Ҝ/v]LeZ@7-0^nNRnVrS |/ >s)i@LlhшnJ4k)Ke{|@2{xyϳ%CΏ,XPPl=Dw SKkjbJw(Z6nWYqEn,s>z-wJZ^tOˋ?pi!1\- kZفi:]KŽ#Yʹa~\M_"\<0IUUGz2g-Rf 6Yf87{ѭx4ԅJƺQk"/|M(oE oIw Ox Y(>[L F3CG9^̝ =t,3yi~}Dj~Y/wrZrLu&y&ɨuq{BI6O?RF`)dV'}{K͢X@:(ҮBII\FE暠k"D%H1B6ڎv2m@Z^CbOFY E#zmfPR$Cd:gkY"(Ǹ92cߊH;9 4%΁I/ [ꘞY uY7wJ3m6BPHP<. 5x1R]){. F X y!qrmY]_i2q:b ڍֽ밇;3@F~g1_bj7u}:@iޱ1A 1+x+ S1I_d$!mH- 13#ቪ{"I1s&pCB]L:k^ɋcRD'&h 2D[DŴe<2/c?RO^Hq*=*^S+nSE )+k@Xg>r4A!,',p&^р͈9VTOA[. !gmBÍk3)ښ9{4݊R qSu M NN_t5g78ͺ1h8u57LJ΁,I((o>[Ahvi#A|T?EmJcQ(¡1pY(39̢ڽmj[c7˓ڵZVڶ=1i<33=?&2K6eHq@sӏ:0kc}y"!WEG,L,5HNuWև٭[~[*kD5bv?J1O,-X]k(3˄ 86hKiX8f8c*Vl,hE T2p+H(-WdUֈ٭:Xg5.y^tbOZe=]yEŞ1bcuGؖէ)ieVǑZȎMz5 }@Ćhrjơ,d2uw>w }pr u6t' 8InU(}IvmL.&ȷP2ʨuB3:T]g4{CپxId]@&ӂ%42vFkO]V/RW'ƱW+T>wg,[[mwkJ;^0* ǫiN_iqv>) E7KˬRFygGص)'|(;~wiW{S ǼGc.1y!Y%o18J*f0tBι{bh4=Cv ƶGZGBZ' AE\mYR0@MHaZNVrѣ砭O =Aϊ=Ųq+Cky6!S̽A\RiD2Emh9G&g=tJ1vַEW_8i5%ʪ2E7ne**;њ6aQqёБ0\ơ- ~+.d"'S,@R]b-k~ ;P"KeȇwЕיbH.'ӛ eB rzE`ןy qXjWQ;Ms2$QVBN|$eYkƉՂwK0Fڠ ](D&-;\0{HBC b*Ancd Zݰv~R?XL~y6Ae;Xgl|?`Q6 {6al@8\Oפ^M'_H/OK`Y_6-4Ow:zR^cҼ?to4V*@IP7<3CN$ÜP9쏌(c4gid㠴6&աQ9D k,|_ٳ64ؐgTp/췶wj_v#2]4bcjSSI"a"!Nّ8cFiZ&eRܩĴ7ږ<(%˚cU  xg8'!9=xC!C *GVxe5rvK8 t28W+_wծs~E;R LI-\uϒ/"\1It!4 x |!g&w#<5iAS jX§* jv/d6kóTL3ޗ%:S` Fѽx5;Z8QczV kz4i4O?%e]ڙJHMlB*BdUWI KVJ@Im]| wNcpj4ஶEc[luG~]$( *9cXA/RuI$sY MJs JRO P*K WR \T NΒAh2t k7~FΞ-|~JfnO.;;3k2 # ˥RֱrάT0NI>,騙\] YC Jl<K7x.4qt=+)UmyYi0 kVaۢWс%g$"yd.`;%${< U}qZmKM'Q/wu^CA "/ Ʊ[ Jqd=z-|gf{n)Ӵ er!#HT0qJh9h2 9J(o2&;S1{Az =%Y;ZTKvg457".Bj?|#\Zɛ+JD"m r42|\a]}\ېN{$ؒ7q'8L׬Pg()Vsv.xgtF7g9)˳4rO0,yE42O̙mdR`\ߖ~+\B>zug1k;nWU$Umn4ڔgqQܮ_m&[v ZQ;-GK~mu_O ?noqLops8ΣŧJw"b߮m?zqGiA1ȟOR']v=Ft>|&qFE0HM!IdW[F4v|W4^Zf:"dç3g1~tG7}v/C\+#{rYKǚui(U6;> }nz:DΠy ˍQ:ioTo^/~z|Oe߽{uGS$/4)vj翶ZR]cӮ9]Vrԓ _/?#?JyI?mڀ8Ews}k'_:Ml?t'(6e;IT+y& 5rs6n\z+xtMlii.gtԡ$Nh㔌>[= b$ӹ⫌$xbd \zmӥC;`pJ[J۩cD[ǬI$)74* x 10_A/ZoSOG| ´;^95CbrˇWwSwt@R ZV82D͵WZ_\~igK4hz9 u*㚙 >?oS.ZtWXMsyw~^~Je?p~P ޳u \ߦ͂ӽG-ϓSFkk,'Aưa7:WXe5*fo Ժ 0[yso\I\1j(Wo~8wlsFO:/)gqx1ԟsZ>epb~ڹ|!lp6Sߏ\9Ld.:aF>UWޟr ^w4sWq.GGYțXO&9Ūٝ2okI}.S -"Ig$s^0#Tz H^HSsF8}ﻡ#a ^H|+=_Ug57eqm-ng5?9rO&\o;iQ_i~6AKCh)* +%},ެ'2;.ziNK徽Z>_uK{d~T|4|;Ű1,ka{Y0Wo !My|p_mn8Nr!/]lK/:] +>owwHy~_—6MbfWB29<{o1]MK蔵\M'5Wf$;& wm:`|Ũ'WQ$.(ǹrQ)kYת]tsKRڋ؊Z,ae=";q6mjg붎Uam)i{Dk#R|['>[B8[Z &d) rsw[Tmڧ7aL4QJQr%6ړ7X7.?NUM1s%( LMScy u1OY92ryz<1OZpW(yU-. \ei9tRB ;v?|ʥ8]eqDgi5=tR:c+M 0s/'IR8Qj|Y1!B 2w]g87.74fv!g!қo=lô7A*nIVk{gd ]xGo,P?z'xR8.轏kQ+ kHAx`.+`3o wџݫ+Iq)L^rƕ[zgwYdUE"lȟ]VvZ+V#rZ#Ņ9wBiчe);9rƑ-5W kj3q=;m&sr`WpEOpSI-ڿBv//|z1Qe.{Ϳ* G4ajT Id1YC4E'0gU0\Gu\d/1/=*!.8{=5)c*H`< r9I[l?6̦F |Tt=}i&gd)Fm}TJBl}Y\jer}[_R>%(_vӳӶJ>ʿK߆8 n\AEIוM[lY}sq޶@R+*K.Tٯ)6tڭ%YGz"\CL$]Ԃ߼3$ T0Q"l֛r(}!αY_;|!6p'g_$( y7kg$s}QjYCec{_L*95n7FwǪڭ,ϥ;-_GݏHsɱp;,LY7$􂍶NB*[)dtd0- 0wSimsw3j ĺRp1p-K+R@*ܙ BDNJxaH.h h‰kZ_jYu…jM"\jeQϳ:b1oRL W!qzRd 6HIX[J A42g=2f)4c[,TPpX8(m|υ|Y/mvyǽol7G-@FYTzl R^Q#քHbBX ģS8vp 1c[Dˆ'DJ}(&)MyЂ3.8"GO ͭxˆXLCĥvjb\tYLKES OJw}Cbڱ%rR'm{ٛ}eG= 2 $T8cL >Q"'L/-8Cv[J1.5.+$T*1(JSR⬧\ޑϞÌ7/4q|y5eesM^NJ?ch̬ks?JyarZ ڝwjz4̓_')H^чlj28JPC*a a4[uæoD}cDX!Vvv:uيQ&ʝB%f@ItI{oHA ~bw1h͠l=N{Z A)G"8CaL(jUS狉"1$6v7_ %_ wZ)NM]bI'ӯw&KZc֙6Ǣ%AyS5D Ny RrT9+)` >7*pPqDvr[]~,dg$4cѓ$r< (#K Hc;sx=Qфm88O:IE@#6q59LZ z'(g32vPuiȶQ8z7Vo{~F=ؓuؓH䈓i@)䞂M6Fu8<PbΩ" 58FgPɷvM:]u͑Ǚ:$q.!1WS'. YgvjHND Ai$ٟcuf %fr`ٿ?'O\Fo =! 5-n5C6YbC[(4i'vOlݶK:ra-ouE.kZKV}6x$5asaW҉>Y}"RZ/T)_guwO~~~}{O0QWu#0.eN\@.snD/Ol`B[N7>TQؑT[_e@`ѹO69WR1x^V)AVY̖Tލ1(NQiXށ_-sd)KV;GJQ&YxX3ݻV]d e 89'0{W8$6JݮOGF݄fZm5u ilyDb'Ed"%&V(VT5P\Brmk;W~UVt:YY~쭸dDno56 v9׃X[6J!'7 ƙARLsL^BEb/ih?88)@h S* / @Yj%0,3Ɖ)yNT` & moli6BOw8*-.[* ލQV[A!*i7y|T|}Qlde՛ENQʔ `K{ dޥaiNE{=d8;`5faoL]+`ד|ꅬjUnaNtހkGTRfrr hlB8"\ mm)8"֊X+"@$M 읐&RiD\ R<$*:@֪F,B0#BXY\жhA #(H8H׬FΚ]=k\[ͷCs7pEZEgW+z=VMppS]+eZ&)z>z`ie9 ^HÚčHW[ԇENcF1Zl@d4wFm6*GT KIw`PQj#g5YA\d1~w}JU7_(s-+=)BNXx^UxyM,GnyΘ$֥{Bdٸ@ 5&SH)rm9XͳSXĭ zh4f‚R/De !I%iKj4vCP q`8-RLp7Ϻn$"g5Yau-p;z2: ZP6E9M8fY1Ǝ`O!H))Jj) #U {ɱ 1q ; 32 X5xmne?Z+ pb%h\2]*""JFc& j^.|1_jxcH;G@)*ݷI#3"`XXKRS|i 7!X/n}M=ګ>\G`; ? jtЊ{ D$lYoH,L(r͙A6ƒE *xIjK6qrv;sANj#rӥAA,uPoH4FBC!z(瘀fmPJ#@Ȱ, Ƹ% ԱEZKFRQ@m䬆SR;z]إ2˱R.K|07;v<;ힼo HQE}Mܧ:Gh6WHCsKFD¿A~%JjKD"8"%xYDac#t ^)Y+’E*R\[ cOPMcɛS"!&"Ŝ(aALu+$;_)XcD횶gݯ+0iO ɗ?,HoFbJ~ؕ3Ek ƈ[!\{A ¼zN|g 2NYq]i5P#9oߤtUlko qM){"c̱|лkF'gmI 9Zr;Afxl,61\iQzvKէ$m4kYVkmET΋17_o$&+k@!K2&tK=sYkBW:_:΂xvyp7 yK7y3bTEH]K&~ˉZG@jV@=A*еC*&$ٱSE=usR%qNp9m=Zh@Mm^Mض\A^ H4!ɝHKYLGڦua+-Á@B$!m`m.)Үnm|w ˺OG11\k؆g_y&u9n !h3@2s=&ske٫ue0N2f1b{= 5 ƂL-յK AHRVa S띱V'ZL"jd)VHKD6rt)eի3= i:ܶ:p_{XmZ@n4k na_q0p'bWzu&\m+v :ainm"/yLB˫N~QS w1aņP:ES,zNE2˵v.͢B(CɼZWlx<)*L5O\Rp.wy5&vOWT\Ŗ؎ULxz(.hOR?F?ElR:Ԧ4F7ߓn6t7mϊ9W)121МQr+\3,eRM]Eгm;M̳qݽNq&FpfO6 }zuY?1褑Nm-C<"Ttyh6ht[ᄡpV_]sb2̔,kᛃ䛭yM='Y# 2|s'>0wAa:U7Ӭb5| ,wÙs8ă1ǘ3uá lό0ŎbiWZ~1-X^ul;i+]EԳԺasˮ&pt>5uȸUU棑r$fI23 )|22aD[b8V cǥk rOG `iJDPUp$Uk#6ʠ_bJˤ Lcԣ,Eɠlzi1{{cRS|˽zO=TcW`ŏF\%r5;qCHJTz}#đqL_N^wPJM7A:|vbONt tf~* @\(P0 -BPrD6|D0FCJLA Z ~D V*U"JTF\Fq%%X Qv4*ȱD-.B4+%Rn6P.>gǪdlZ-.DK29=G4~|3߆ə|a3$sDN6'_ooF1J QhD-=D%kN_D@5Ho^CZ(ڜ-rm|̓1kb}'.WǛ~5 ܄+s._vMS<{;Tӛmc59͑BtkH mʂ=K<+nq_/c`bҹVpb#Vՠ u=]XG1`-@KZGK!hL8P=x:FY^1Tk%i ֪k=]qZ9r񞴃dm}+davk k<Ɨ´ 'r n] !L}n%&9);&)ptڽΤ:mmF_v,Zʇ7{nzou-a05` `0Cb 猏Z:ݭ[r z=<F~Nw9buՄNH 8*bI.8F$psj(6p(F f1Yw͛H6PHllyRIt"z)TJ F"OQlo}e:?鸎;YZFgj&>E=Bc)lkl}%RT!+h Ռ*D;#bgAUOcd J&U%mtsf<.wՅq.]9ϊ*t0;0_|`*{{vg_fAjQuu-Y*vى# -zoE5K:HqDࠫMGO&Qs`N8Xc#fMq6Z{wUkZ;m^kPE疳0J{nr)DcloPBL֫zٵ6⣨Qu^(T%TjTT`}8ܨ_֥qmlQqW5"n4F#^Q 7EYAqQY=tՉVY BCՀMV~MW۸l^<& Ub-hkȬ4J{<`85T/Ζ9K`7zqx"c!j"d"5mY j4oF-ȲыE;Cg}>|vCEy7Y?mh}E<*Qqp' L$?<rYvwpã:"X{ }LY\x@%*;7sRjoȚ]׳1S(Ω`.p=-zE9b6xjȘQ zҳ)uϦ0l6u/թ~Sk,]%t:ϩN n=]B0LM@P_4nMڄ56a9xjNXG]jM#IyKJbz%{GMԊꅚf D#m\c G͹!ec2@O^ha"SHuY&޽?M]QW3wBwy=sKO>Ouitb 1W 39=كnuI6V̘*(B6R +pXuh\ 6;Nr'E5G'9{a*b%J5c|T @9޳2Vr?Q*gHq l2r{By2 2`K)*Dޠ`kIA55gJvl9\ٜQ^Ǩz1dV[M:xWzoV6R- eCL6kDN;$~":*['࣢)oQPrYĒ>;ւ/As`25 jNIg»uX { <ʼnâ&Q1P-"BbcH=QCǬK0cyO ߌoeEmͱ%(Q %ǘ)XWrU9[n_Mצ0· ~s$D+g'"prcp*(x30Vq]cp90F'ɬo\VgGGeR;=.]*7=@n 6mc& ڎ^\uPJ% S\ɓ5OXN6P)c AP ܷ9Hju88|beHAJ <'Ih}ΘUhEE$F \ >W ճa~[VϛPmުnmT[l9a U(VBb݋졖ݰ>oK`tՓbA ~LָEQ O$p6%؄ zc(F>S/>!su@g:q؎_vNv ^5mGm;:x?ORᨩ&/C]f_>ֿ_eyg{:'Oȳ"z1_5-1НSgѳ}hߦu]f>lώU n{y:i+g g}lLz$l+ȭtAv/wo1WtZ9Wn0rX9yfyX\G U;އB>nɩ\9*G]Nrը+ylYoKeqLlu m>c9_Ш</noᎲӫ7߿ﯾ{Ûyחo^:8n !7 >ރO?~dF -vo1t>]kuF,qts37hwlo s~>ZǠӟ~q\Ԉh8׭'JE"lO1r#Q\bU~i H5hwu85lY].O`'0?t߬:b wT69T Y?*mԁId EW3g"@ it}V_9`)\1;k&tʒ'PveV.ldPX b)bP<9B 9Zxa<+(V]tn⭮9V˗_\0y rIa28yoӬ'>x -9 Wݥb2н{f 69bT4 % S/KқUJ =3:dۥbؗ4Rlo,R~5e;xh]XMdɃ0rn9> 8#gQwG瓽-AU+tq3{c u P2UTiZM(y}S/v_' &,3*I)+MI B*7f6Wo٫9F&& ̎l$e$SYR $ljޅ'tGŒ `K^*E}[^`"7SB({,(!͂ j0I)բsPW;պ~2f23Q䰟" b^.Z 2[)FIƎ"OfSI dv* u}{3h|"w`Çwl*⻲{agDb8|14Q!w5{&>bocG7 j:N$ .rnX"c7HȎJVf \HP<( ](^RsEbIRL v #f8uY }G+tħrXckyg;; ܖzp۫7 y"ѡy0s᥼҈ k&IM72Ek O-AM@1eBl(s?{׶HdEY%x+݋6;=.eI#v 'Jɒ-b;v) O(&#yYT&-X:˲yZATA2 E*XFj%$CMP3@p&e6Gb% Xsv eQY?8BW'l_cs~">yS1.N EBKAM8Yi,Ie "A2P"1aB0݄xoɡi Xm%tp m37J)t0ҋZ G1 B~[nl# V|UՔ|?>aWtwA ֔ pn2g;LZ GYZ))&eplY k;C/㙫&-Anqs^xm6ECe?x<fr{xm:8ԾvO/8RgZʯn5֗?:3) : ^ˮ~7/^| ן:8;Ѽ9fY'տ^5 3B׏1SfX࡚0&ӘsUbUIPy0]9dYf3#zKj}zŹ $*sHҬgڽt4/_R\ӘX}}O}/󇋝*5;,zkOɵ{W'^<7嶯S=H ya~ +҅^OġZ M_Ԏoa]Nmx0ʫȄ>BB2u*!界Zъ+LBd". J1RR zX'E(ňDӄ  RY^ RLiqy7a΃-i]ՒwE_p\USenxJM@jt>-<ĤYH**ݰq^8 F*F*4T9(sAJ} /[!VB 1c:$oN)"8 Rѐ%@D&0R D_;nKQ4NScp@ :g2埲D5*ќۦo1Y^kΞ='zr~rݾO[wsk:=*oĭo8Ļ*7v!Ⱥ ivG,Ճ᤮{mc6n֦6G4@h{չ[=>ett5-zxG2ۣ^mͺB,YaɪW{Txw `JnYKps}O>=>c~O*~'޺F oVvC4+O|hVtxSSbtw秐zZq!rkc3kokSew%+!` R9)GQJ T6R䍠m9=ibtUMv+浙=0ԤA3| 3?f%`2)G B7.͆FCϘLmn]ajs<{mѷr|Iq`|Yk7\K5Ȫ=כWz[6FP8խ6-,{usjV9:)߭e0[r.@r>weKIg`kܠ԰f m2,g5t4<=쓹]@yc(j8x -Xvy|lߝʣD6%{& ]iG4H'~'M0"@%rZhG[="zy|>x|~\Ch eD۫c7=Xd)SUzT 8wHq"qAQFj!HʂRV0H9)VӼ6F%V~8dHChڃ*bYX*ebYX*ZPM%J͛IՄ(_zZJڴ(0- Ք Sj*y5ӇFZFeUNTU8_9 "&h$ˎ% X]P, ŒgQ^K5 idC<"f/ͫ'mpI% M U+AjgV3n !{©,p;)y]Vtt2F !;m<8r(Cd(_]P`H¶R9{j< \he ԡԑ%ݩ=pgԁ7 5Rbؽh`|a̅nBX1`4Z#BiNSlcN3,:G@4*jg:#vevUj]=àkIS6%gL 88MKqKhwJ?J'Xkcdc3Rk[;Mqbʻ[݋3^h,ĒjOH2?Z'yr\Yy^_|-8"ӚW`u" :sHFqWSnٷng5?.CbVBa4lovu~3~ڏt~z뎊w"h,<R|hVt_f}rl* ȭP~cC#}bHF%+_wwrFd9hJp̅yVb`i&& Q̄,:sk@hl %p) bTclAL&ʌdXGhk1W+5gՊ<%-[-[%{(gQ׏b)mѷrI jP#ZHISaJ*c`zc 8cUdG8GcƗT>c`^:+A`xB?͹|\Zz 6jAsC rp۽[P29J(y'vI/PFd>ʩ:B LS O!Ἄ@ɷ&_C>6;ϣ``M#dIEmXDR#8Riux^7+ve>xVX*D2P [D"4褙zrJSrN8ib̧YE]ƺ -6 3{`hW^JGv]3,J8lMUIX[7^ʍl^$8#oJ mP+ʷ͚ r]p,-O:Q[cjQASMj3ڋ^]}^ Bx:LRv='yVŋNY0Gr#J__. L)''vȷ= Hh|N&ה&1AR-a<%ۊ$K4oc^yN~|%Ӧxτ֡uL!12:V^#M(}ktzE^>T8ƹQcIƒ Y鲓G}7[6謔=v[^s]bH)z~=uS^BÊ=:x2zߍ|B }nuT^X`le (huu\/}qIb$0J(ZsJ*.>2" 3$L$LZ{-,bH]Ou42QǸָDphiJ#Va[9k|Iuxpr7AY^X=yv4:ض}j`C+FPuLALWτW4@ ŒC&N?)"Qe8kdeVy9æ@+QN]oWs̭.N]îi^4/N-:RZIu4; eۛ}T {7~Hn­)DA5"T'r™*1T-zWz#e#FIk#7e$;`19K/E[4e%?H\zV|E5!\UNq%d:gU chXS`9+Y듐I?iu%Xp ~sB,| J.iLg" 87bʮ}5y~LSqF~Ƌr?-w_H>>!u6q?r!ͫR@#{Qi8rOpJrRf$(d%#)VZ;u%}t96 -* &D2_NAxt6 7׷lK7in"ޕ(ad~yk\]';oZ`wFLgS'4W.^}6ހH޽x@8KRKdw;yU|ß>4\Ո`= FegoSS{fC`Q{x+^HLT\{{t Dkf.wvoI,%pCN9|a 1RXRDX0( a1fb$^52 N0g7$E$y-GT @qt 2#Ӫ]DݸeaVՌlN)~HYrAc) 9+o1n\"֌ٮ7Gх8㮺Tօׅ-fc:+XkNu7w7dw ~>t~c:' ".Qss s2@V tvi{X,W> &CR)F.Lr3*kjla4L+hZqWm+kmkn٘42ѹP";)rqH縧uU}>R\LȐ+DK&!F #Ct]!eX$a5r֩_$)`|ZqW5"׈`9RKK. ,rAi9y\pƌUVl,* )DPhp+b$OZ(+*pk]#~z"G!uV㒻EWY/^/zqcP 8"yqabALL0T<`r{xzPa5>>TؖĿ]2Fn] |AmNЅHA;Dۏ-aA ߓ'+.~:?q=y Q"A3i ^*А ')vE~qwi(9^}DIhfY n栴6"HUj9:?J/k,|Z}yCGRC|c"|xu3t?z%q}9Dxj)ىKޔ4-1YȒio-Be 9ukFu387yIex@r@2D=VPJڻ$.⬯@ǣ4ѸdΪ͡ݧ<~EY>g)9gI7X>#cPc:FDŽo^rInCf\~u3RcdYi$2#H .:_Ab)%zW6ʙgK+0qbdYjDV @kD&g(wU#gG9k_- m`RBdi'Sc8^R9)24;љ2(p*!gI2D"!݃z=!48d+=/O .e+Q9)sAK9Hɀc`=Ԫz?i)ELV< /3COH$R'K!M9'LkQb⨢8k5cJvy{Ax'OdPDy'v7Xe 6iK0 Vk)'ۛ+"DlM)4N/`[Tc ,|m]8%#۬ &*ϰe /qK,ZKA cY+$P5{'K0vPq]r&,E٤&MۤYxۥkCRZ QO }#(P~"ث?+`~hShQ*| w4N(&M]m13LsdQ):cXA.rTu(Jxԁ9N@L,rʢҘ1E&DKi="%,M\ 1pQMB%%ʒc9jxb燴Ԭ ѻgk^˵-cKcYj$}fr}Bt1d (̉xv$Wx.4|6)o(Eh~&Ea kQYVE-pUtt`%@OF"VU/ǖJ-wbWä-;zMXwJtB%|y8b4"@H(bJJqd=DZxgGa0PiA"B6^GI( (AL&?gTBy+$6 >5/}V{>ֻƒs}!QLAavvaů͕smeENOm2rHd<_ >~(0L>8)k0߻ϗ\wڵLZ^}F7py_7ɲ5)1r89=;k2/dJJ]3?^gF0x,u/!\.iBi{ qp|ͯi SR^>Vgfy&:9y6lRS2!6qܩ Gs壍3Kj9G_מiz5/ˌg\Kwe{‹Up91:wﻹjnWՆ?Ϧmp98|}$ڑ}uÈÀkYe$ VG Y,|4<\Lj̳醇正 yCu^Z:)jq!#e`M8g_yPVğc>un?cS6tj'lijןޞ⇗?Ϸߟ__~S.ş^}=Y L1kHpBM/{|6y-54Zah'j>|7+pQ׌qWħuhҶ$寃g~~8Ēfr M+I^Y]E] -Bd_t?&87dTŋԪ q7 n]Dz%[ZF:o|o71[S(gey\|Om>Rϊdb)2:Z'v~'>(ߜmN0.f̳dQ@028tHE`/R}wzud2L㕳 oY@c^IZ%7QhcfA9iJ  謣 @H>H7lBTr5Jcn 8-]ϭ]+uuEF&%/@hsץ09&2_uC/i:-.lOٲq6eC-e1wv~U-qz^iJ-nۼz]x~4̻m.ehiۯ eܣy7l.Ysn›j~'=xWo|Is}G0\˞<@|qЭPq4N oyMGAkO2M]tr/_H?!xZG8o^+E*1\m)EKa(+ FRn5rӺG'S|s qhP\Hz/ouݻMBM{vt/$/E[2wnk:.~ Ow޴E,^+:Tp^ ZFPx*+oF1|:qEW#I+;`GwW0>'{6nq Q-qM)(%Q-q35=U_]j0+ >XncrJ5t)]e9*KSX,$X KTh).pV+ ".#߄(ˌ|<ٮ2]櫵Agܾ^2Y⦯pa]VA) 5gᕂ`e*a`f',|\)%ٕsI y:27Yiݰ_3@[@lSϨ:5g0[Y,.qgݩKWHr>r47O<%"F*q5"1g/>b3G*[Lf7PdBUx'sݷYiqw<] Y(>qLmsseM3g˘@8{ЫzEc>n*e5ر*ܪXܣUqE^XϹzsG^O3Q&UI~jW MWz~ؾ4hW $l&c8W7.DZY]< .Q(墀ي,7k*(n2ٮ%Q(Q(qRaTɥ\K đut:yR'BR^"7N Ve[Mm[\'v`WJuÛ}T`]=MKЏM,U_0iͤ2*>_Z>@UKIC(ܫ޻ߤHo풖^vO}zt.;׮V7-}/F?RI t^Ҍ]; .EA9~ݫǓ5Em~-Iޣݺg{Cgm.gFʋKԯֽ[jȯL_l{DL8ԃ1-2d V>jX TJ8}4XΨȩ rkBg]뺮XϞpy<`:&1 QTB霹VY&P΁9\׺qZ_dHGA\#f9\:/n>lnzOD0UO6SY3#UcKX`\&X.jINA@^@vgy"m K?[; w1 p9 kr>(9cJoɡU=B5;=-5CZ0sڙOxPJ+xFGC J-#H:bݍ*`PH1R Cv"c敛*f`n3HYSB/"$'&ek7%톊Vj}U'=Tzz]YᬊʷgEn֨0ƙ<LJ%- 9 1Lihrx%=KdM<&YyĸY'bmd&XTj+Xhz,|r DӳŒt7;y;h1jrF.Qs\O2;' dI y:S:y4M Kjg'dBhT0*Y1pY(ʙ1h2bWvqZq1kiǮm+Qg}6&gFL\fiYf}I8`\(-9.emt]C&fyQ%IȢHq$|HY'ɩF}e<&x:T{ҏ]*#"#22XRxK5$3˼$fh<kU]gXE1j%rLH!BG1JɓpK[LZ܎_O^u \-Zg5-]e\t=.k`JF(շ YS. VӎPx8ޝ ʘ]vu5YlsڜD܅QA[яO .JK8oIOQt ޢ>hٜRn.]SL?;=7vߦL}z?S%#μ)'2iZb%*2퍶e !Iɲf|:t< Q@FϭBG#=Fo5qny7`4 88l.w>E ޯžf!LsP ;LA3c jl@g(Bd"ңbr=*ΣڣNS2`Af1l4`Rc2^er0R 1愰q6Jmio)b5q# #F)/LE!9mt~+|fr6m5Xtfҟ/oP*JC;fI1% rO~crAL03-6j7gB #J q̐yFFIpRRH\]jge/8I*U ,Y 'uL$L %RP2de;&Ζv~fAQKG,җdI!EyH1D9FsU 9ND${i_I=[Vؔ␭P|ӏvpm&j<@%bA3eT)H%/uG@ZBtGD",*S]vnBD \4p%EHD#jz-|yxGMiv˄ڔl郸Ytoz4 ]>?4]PVs0Le~m~펼m>;=f]1q OblWH? bߜrpG8~NO۵wR;}Zw6 ۬*{2ĄIQ4eL>=tB`xuw]EwZ\H&d˥f_yPVLk8V.܆҉uÃ8|B7?74FcFyI{P\i[wjܚŭx_y5y}]0.ր0pK6hS6.fM򊇅tAZ|&3ԕq~V\JҌ*~J-q݈un]r@>e6qn̯rBIi/Zp +q#GEO3N ^ֽ4v3qYX%J*݃Luu+V2H&"hodb`VFfA[ZcM>hto$:`yp6d4̱D JR_# s%N;udg&t:olE@[dYC9U .d U!)2(o{aC*Pg\mN|'/9zUxYi?^?,>Ʒ W\KhΘɦY7^D#+\^zEbg';@jTr-|0ك ڟCȥXfdEY"hDY@D'ph,qdz([>o[" ٨hXWIΓ9T5DWovW>r1gkrlOz%uq-z2 ں9vDHY&'ՏeBjP}05(P* A%dBԚ,1%%?T|%?ˆ*ͣ<23hs*JT :2$S*>11gl!ʤsUv(H7! ºYҗ]f*cV2 'd(4*jlg>'z\)ulM>2o!6M|J>:y"ixsj^R$&K/D > sROg ].bteEĘTYLz/sH Y8 yÕ[erFURu m=Z"g9d }_z4 /,%7~ryVXEr΄.Z$Yvqlegċ5ME_M/p'b#][BMַ rX~Xgѵ%6b}WZWެ+ n,^j尿LLo |>SoVvWNz6q.^l}x)x__/ů^ui94 q0jT ґ 0F4oаL{JpB0ϰ:ϰ:ϰ:0GLFzgDi2 e9g6[k `wJ23WH(a_H̼nV튪 yB܄ٟMx&ijT9xU>^o/?\%}c2 &'/53)Y;N6p 5zPI縬Ay`e׏%ʊ4M4):(ף1* P%&yfpFHR\fk%d܅$8Iaxtf/=K~YC;F" [癦Md&Yl?uGɥ rv:ZW)h%W.)Lz@-i7[U}'i֔|dITj+mߜH 8?~jHCv3^6?:"lA1 ,6VKl1dNvz>zz 5ZvgPeOٮu&&~j *2\ۧt>b >J x,vRI'h%Sk4Wj4ϓh4ϒR3u^xF#:楯3L,=᪈ UVs"幅wpuRxSX9<뜎Wt=jte /f}h9d\T -#/WlOy>\-68dFIrLJU98;/aw|uW^'po|ڼ7߶hIhxkk\ߡ%[d"tmůynz fIvHvrhQd͍+^.~6w$ 8O1Bj>iZos8 AoͶCjPf>[wv>Lv<ƣцW7 {]&!r5H̐lȒL%c53aVBiޞK@"*2cʂe*s BrF$n"S2 ѐ:.4ݔɢ^fQ]AoNV5KhpI %~,sq}lZr?RI:2$-` (R6("%xzw%f :D8"< 7 NhJ O@N`[$IdBu%#,g(}sJ͡C@\?y(˄|t;ӕ'#DpqQA .IȤxI3m˴#&COO~k}P.ܠUgnbYWؕ eEl`X]ic,R663 p>^$2F% {HĹI,'sv.?Vg2Y eNx;g)Q3F3p>U*W3RɠhfeH6`d^%{E83RYhP2mIS+Zن$Y)&f&T^پPոxDKqڳsR͓׮VZW/k߽"pI2 cJ {38Gk3_ b`hCcʘL0Rl=DNĸpr62Vg;2*հx)X:,<'bz[3'&a:Y d#vR(9$ IΊ,<ziуt,iꞲ#U?eE`{P`S(zQ:nnj̀6U9̂68ۏy6]kvuejw vƄrLA!YYJeiۤcpbPB$>LupY*ˑ4dbI"išXP)"Q1)I;W[~eQltZ㥈h*#Cčk "s48!Kp҈(2 8M혃R&Í3cXW34n$ n\D )(k+m#9_?I]}U0ǻo @CSRLZ}wCHؒhk]ٚ{Uuu@4ȲOe(:+kjܭOjOzqߦXg5.^bo=k|'Q5'7[:AȀ%}60 ]B/n>ԇՇb}*bwbq[\Dُճ[SBbpJh}75xk;"mΩI] yJTv [!nE#q\\g] $DPXQ9GKֺ:;cيwy➰[(Nk:sBEg}N* u YE-jkZS4t1]epx'c@3c#|Mی^vz(Uo|)htY*!|>6Ri|J+&46Fn,aצt9ot]t츥:}xndէӓNxUڂy=t*I;'9LĸYd$п6jQ CFΣD]+y):ȅ)>_༵AaluO؀88=H+,BLmpg0iZV gW?7}OGBoݗoGOE457tc Qr[6oC;WƶGZO 8zdôDY,rinPy)s6ޯTNd7l㕷MU=PuW)} X% )52M::rာâ9Hr=t4T& FH`IyG5<3үj|g{͜[9"s.1h0V}rJif Yˈ9D: |> $LYp1aX[gV12:fw9?mpܠ4Yۻ鴓-AM+tظG}#S{B)!K##4z%R4EDޟVJVn,W&gukkw2&Cg',cqcTa+jF|#x2rw~𥉉睵zי}!n,ƳyŢu^Y̽9,I/)'dl٠|EYZy'MߟdZC#f}px=q}~fr.6.EKf(+ | dZ9k %rV.k/.ۋ?/*f#Ѕb&ҋ>V7ҽ$te_2"{wN mw>Oj7/sչ`wPd2ΕK`w>!N3dR{zJL΀qp6>H%\d@y ^UtbSlܺiIUF:5R*#@q=d1u-jZE}gp?~/&y+`YلQl"]Ry dƛ)y Q;EZ+{Id}Bod Ir1/gO6ࡶG[5";d_3QBCC6WGiij Mhv@Ig)XR/*FqEk2R!LbΞkLjC^xau %22֓fIp/g,"F @ ,KHJ;Kr#è<`]\ K xTU9s,;a@#'4J2  LJL0cJ@O9+*Y5r֔V~lFۏ!Kkm • SGg[2r‚1q C$z)Pk'%5)1XA. mY[pB @K9Q#UjUZa6OZ vdAzC2D&` &2P.-;@{ŝRk9+*cf'RbOܳHvLlFkyj:ӂ$4]_sJDS悮S9t͈v2Iק^">YYY%xq(J2Z6Yiu䟥ItY`D(Dʅ+$PTZQOlўGH7Y$Hy,s/KF`BalHFg!{&IruѹA(,&K(+)6HСB%(kȠhhYݗ9k[6vKN~Q".4esf:RŒ&/AXz]2Gt Ó)A+hfu㼓1 Q"iL[Tt2*X4䌒D$GPEpnq_*-Cimx8"ԫhjRZφi0+?/D\vL5aZ1O=EN$&1/b`ޅ" cҟ^tC.;m)Ŗwo"O0apř,i95%w83PS"g-aMv3RG;ihŽߨYr8+K40KjL`S8_3Zð(W(f|0v1ŁiǮU$U}b9/ -.5n9t)z\~MS(&dBdJ#͇9-/aF 2dv"8Neχ<aLϖ۠Ot:glQGlA ,oWdWTت$ q? }V.h|tM|YsA󫘭fv/-pn@ v}6T$Ed10"C*  OD:`yp6ḏh"`yc1(t.;uY-<&J|lE0PsdCgb0'̼BgC.ר #!y/߄ UxWVW[ue% [=MXwTIR&@l4 )XlQ @S}o[\IQ}gm3h;x:vR=R( Ty!%5,Li脆7@Z#5a%KDcJ DAa6ڊ *3h 1*8H:ŴN\L:oYB@a&W 8\֥ȲRYre&=/~BR {00r9wHC746 ϮĒoi)7mJ.p.:!ALͅzS\ w;W:7I.>0ʫFJ ZJ^Ȕ1[~n}/Nj\+Z1Ƥ&g{C(U`1'#n+*mt3 3V!rcJL@I|T]=Iy+Z܍Ia8H♟ψ>h49KMF$їtS"ɍe,DJxr9i ^ l y\#"&ȪÚodQ^K.7sʜ!&BܝWvB`鐘S]B$LɅLO!e ): 7[U!1\mi.c̆ ,˙%hx)s:x&y^ƶI&ieߟCY{!Rn@[ܽA}8%-UDͧ_A}F9„DH|㝇eI.xIw7I,Zy6 n<۬\ʄ|ZuBqU6M4xrޠx+X m*۬W7мsJNgnJt6SZ:ۨ; zLPf#C'>(AK/eƫͧc"q DFs=AA|4ڢtR$e9.XAm$ЪC@0H7f _zhX5i-Rʖlɒ-2"qdxnթso=(1ʼn:xC!^)yP%Tc$,=A9*k/%%\2N,s$;:Ep8gGT1Pd; ghʱ>_[LW~(!t-CUL"Fhݕ3Ek ƈ[!\{m=H睩N]wV"k$Ov[2x-fEŮl+Zo[8 Dao#0=U若LN$ }@M`gF8+~=+m$o~+1o"TE.-\Wv?rhW&ӷ+^~?a?,Pʙa @&d- ~V +p? WGgof\(7Å4%g)޺:KRmN[WWVZ̷ý6^ nݲ\OņNyov?`nΨ LoII=xAO;vShI1: Sw;;wx59t(L'۳M1ǘ6`&CW .MVUBZ:EXF+-2_wJp5m ]Zd*ļ+)1At`CWWI=JJ(m RKuU U')uV͎RuUbuEnq"`uqC/߽iCm8gpE=,ι +O? I$D8m M+i 5aw/]4X4}45bXU>fPwCWlKӳs&WWvw+vCqj7ftv+sM畑jN,Ppc: REϽ-l;<-b8E Xg+IS3UPj 'ј 4\2VwJ(jjpU,ec*U)th5J( miIn]`Pc誀Jh ;]% "]j]`Fxc*)tвګN$7Lm ]%1v@PcW %&-] ]) XU{}i?]%Ztut%% p= "% l_nVMRNpEcfD-4POuq7eT31Co/BC?o*\WZ}\ ɴ&ZepWdpk*RT0$yA/eeEǸ.د`VDz &cehw(׏%R$f)P׎7B7U4f;dY7N+=ޞQ+e0@ٌ-2m|&zG{YVZe-C.WޜY$d"iY"Mk1,c`bҙV9+oe&uXo :ri'i(k<z45=LS Jf7ށcɺʺЕP߁= %MO,n &1J]wL(5ib-h]QS+i ]ZE*~+F(q^;9iKLKhF~7NO8U/U֝J:HWHDW 0+^tJ\wJ(mJRDHbW ⍡W7UBIDKW'HW[60FqW]% JƮNgd؁Ps8 ۗ[S'`U2Nh;M'4дQ?:]cm7S =]R{]kz̰FMXf!Jh);]%tut%A~NxR'wFftx>j 9ThA(8+nF>ʄnaimg`*g"7 w9>}NxI+7}_~P'EPgh&LGk3-%M龽I?˜9`?5(L"U/??'R]pR)Iᦛ7< pxnVxmh!Jeq}x3ZHox0k~fӏ㛰s {:f71kV+T[9q?{.X0"H4!ɝHKYT# ) kE1w)1_f`nݲsA Lgri= M(QmGs+:2,"1rbrX+_qdLGEKY?d84 0н)4٥IwAwм%v|mpqyaNJط:-ƮӅ 7R2%Z\kI`w'RtWxsxBQ?TdĎW*:v7o?p]¼H5I11/^UV[3GPt`yt/.ޢilE߾}:[ZiUxdy?uh8^g&I C9<9Sd ӻKU&2Ϣ6.y?K;IҟLb4IMjnP~B:_8ҽad´s ()·])gL)eÿ>_bK{`6˰zVGJx!YgI)m*P@o/)8CvHlc, Vk y>t"n-.PŸhU T/:7 #p$Jmc}J!T'(XQ x>^<L`lF*[#݈uc%T "AQog?~\Sn~E7dn-m}F@8F&NZkeF9(ց,+ct} (JM) :U1kw'R5}`p+$HSjSx՘# $tܩ 1?8lr#5xQJ ٜB7)TH=6|~E N,RDefX elKaGyߪ6ɇkjlx'R-(xvM5*CzHmojRz\T4#!ndQQ0\W0ųs fJf]5?^0|A6ty"||)Kgbnw:OC%<_qm~^ ê+*,an j;[#zHhe{"S nS,_B#σR\.Z42POH0"J[Iu$I‡?Nfx}A!i̘8Lr!K/th8bca9|QQ==G R,C1b{=!ȥsu"=Vꃆd&#QHYu2LwZcaS@`b $g%%7˛wBܰ4{,Hb)ME /̭!Dr M:K`\^䙤l}"(}hZ Jg(IB~>$:z&4Eh:ESLxP.<Yd)3HZ>Vr^&~:-r|=wߚ,6m>v|,i>7: {~dʢi΋?On Sy͙||}qX/=g?cW=7?oN]O\7}=mm=E}y2/; M־F0Ka$H38"CE`$"@B(NОˆ^FcD2Y+g"RSFDD b FрG!eLD221eq Q8K/)\=GW.(.㾛,ͽ&mʶ,_@ ;FCT$4FOb6Ym5i)p"׉dK 'C )Jǐ1Oq3j6?{WFr /2Rw8>xd/kR-)H-IS=3)Y(%QT=Twh0+ >XncrJ5t)*qӀ#}gR^y+rbqzzrelj~C&e\dDL4JbZM2Mdx"(2xt8Jy;r.+g"I&79-? =TAB}?{;,|O@- ޢ\:c"'&dgXHԼL",D̃W3Fejf75C!C u'RT۳; \J|Ot;a*WH6g.Q4ruijN$gfІhyq1m#`$F|tqh4Sq;6(Ϧ'5Zj|K1p< ٗo` [: … ~|zDdYsɲ"p p~>LW|ߦzeͶG;b~#?I.O00:Qkt+-؟?Rkm%sMwi]`g<\LwoJ IJt}vow-ϮOT]l'>H}2_bmA蜴yKNMWI0KYw1xEnlj.gkϰwts6*m-5"c'Ka3yU(㑡 e,?w,\}2*{|Ǔś- |ΐUB;w9gJ]r0gά1s'pW|qqvzOZM7.W[W3}I99m+ߺnC]^Xˉ+RI6mtH~4a@i-DߘϧT!&7^9߸S =Ӑ!@n3a2r;Q XqZPZ.K+*T2ʨuBsrT.{=q$!q]dɻW:xӂa"UVghwɵtIV"}TB{f]XS3M 7+EJbD|U fT':E%*5nd0ʺ)ehwq]sBwRZE_Ca[ ߬vC+16ϞYv`ZOĴR+k3-2tKEooQum@rT59tFEN5oܨ\tpMZDX׷ WGz\#yIFc*epI%$Ιke7Z:V4p:\[hVG(qU't/Ͻ+Frt|W]?ln~zOE7_[cI e0i=+aRy]Iv6/hzA~ѝ;?i(J_d:;)_Jf '.78IJd#r5pddI܅fHY;O]x1}в"ЄwHWI0<:v^f6:n/KAǾN3, 6h^ܮۯu]2뫅 /jyt!2}gq6;}I7Q~6F`oD^coD8:.D=M?.F= fCS?n>7x8V5"XfejQY Y{ܡ)v'e+Y1&"n9d1z-T璦8gs<[hK׬Ѝv75C-&4Ms $O6P yFGC J-#H:bOUb6c83EeUCE,.n3H_ < d$VSAlĹP͌h]m gVTY#{k>ZUQmYYu8'If=>gԃ6 X ؜R :䂠c ) 9+os]EĹqjXXM3BS ̀^n05y {q1Ə~lĎQ4'3rz9e M]E͓יɓnfY2Q8{> &CRؔFQ l&ߎB9T䈙E 8#vNҊ3]K;Em[gcxfDte6eݡMї(ƅ"~Xw.kc}rE21Cȋ-aMB@G#a}C:HN5Z+a5qn֩_V)3>}-2"€"lEbTZ /uidfRȷ+NΑ>m]ynyЅQ1 OmE_nG}iR=oϓǓe I+oRهP78 cApK2;IH`x,m`XfJk,TJCƒg/2$D%YaBB,YG6 gW=H䐘 U0 r|lNg/;x4sޢ}ئ!9n]IL?[ulT>)"Ea+ֳ܆Xͬ;,x}RcdYi$2# .:_bK +[U̳R8)2rP%2pZ$Ad 2 Z"%NVjlhg]w`f~~7(*~ [h1rI/}iAƜRdLcN8X$MDr0&iC{X2ĦlErqlsec;*'%p.IR2 @Q~']n(aQH煗'4I$R'K)!MI&h(8Ny[dHw/ |;S3QF7$I VM`>qH++"$ W^T"l|mM8%#۬ &*ϰd2ޗ%ѠS`уy3 2SuiȶQN?"ټLяקy)M{Jx9vڐHe@l4&6Y{0_R ڈ6&HkpGt RM @asUe8ImFa%v)VoBxU<ԇgK72ǁ8EcSwh\=?I'{8vMl7^fF0x,]17Ӏǃ6!qrxP'">eMsRG~hd_-"˗Տ a){q_hp{h3抣R$tvJѥgueF][w]{*9V֞/Ǔ<>>=r9W}G.+u҉.v</޾ӻ^}㛃^mnF/ed4\WwNn7ulj7)^eiRGJqt[߷1C %=N,ۜLϠ@Ьo.R(7`-_ܺ uka[[Q/cFOО~?'۳y(E۠Ot4kW|VHWIWHl2]ϳ(,.J74# b !Hb7"i]\~-^:x\7sG7󛜭Щf2:E .wdn~o7m}tT2U!J_o\?ME9/۩8/ 9M赌6XG_bP\p)D%={& wr2A- 4ʋ!JÜ:%\̃4^*ha&T╫gVVxeu^&JJU㭖{FouN/r,i:Y7^TH.wRzuX'zWOQEd/$̝^\E\T%3ZYA%8S4ID2wT5Df/(A6Zj<\ jǏr-+tq`G/ĎdQ o|K"Dtߊ%V ڟ`r?_/33a:O`8%bO\~i]:'NՒ]i&IrOv&Ί'GKN:nu͂m5}P&?&|rHũs;gOWR9[M-6,s~>j]u?Y^rGG]sl2n0zJeN8{#:"ͫijz9[z>vV?++ \q>RJ nWEH̐'wSA!WPG_DFퟯ[zNor>?Z,>5mҬ=-ZuV+E`g`ha֨la/g _H`UWC"b}"f\}pZ)vHpU"V \ZUW_"\4mlQH]ߥZr;YiCVq0bD&G^GlzѝmipQ g'$!7` #Ǿ8wIYnXp.j_"Ch:IH;Lga$Bq\|y.iᙞTy>3G *(ۜt^|Gn=Z3:msdM2%9ߌ_h%,B\=*Wo_h7Y$рdFAS pjA6|'kY5Y# 8bngdq}_}=Fv9d5*'@x8Wޒ9+oe~ulj+_x;q#8)Ƒv13ÁN`e:q8 ꤵpU~y\vzn$1+X v0pU}]I pU\ۉ +fNiU s7G% OȫM!^ooxMNB6 kֹ BҚ7f+vLCgURZdO_,,qS:ñl ;6Msg@eȷ.S;3jpp1tF};ĽŃ ^p^QBٴ[ٴdnOAmң+ߩHoR^f>?~uRck^nǪ]Z[qlO~n4.MFt^ٰH%ht[,i軱e~28P#{rVU{Tlo̝jM 7Ϫ(ү'ݸ=6^Kҽ=ٹeGf>iqK.1]6f] b] WG:E0e1BthLxȦs ֢9\׺q\(et;OSץ,!N^w?J1лk[v[{! E>\ŎZq&g 7p:AEʎxEYAcqM>K{B}*E|/%K}P6;,cL0Epd.y6r=ZϤR* d۔.} {HxetVymЕ+UgBvj- lP;TZMs)$/2{83 cKl3VQ\d(4 ǘ<$d$FW0.YK 5Z#Zf'uvE/ɫɎɅ[|q\̮ڎ!_a1wɋ6į7/89/Uqy* nLSQ Ayd3xI"& 9F5HZ/RUdhr{VV5b̒ 021[/Ksg1.- \&yqjXXM3vB] O+ʳ]\fn2 W(|Nhx<8>_|R9*N.g2)nC$2KOE+TPёWEl2IsQ0齐&h6LCR>Y0X0b$AZڱ+jʨm>1-bdmfIU"dbpf\wãJu.r /:dkbbADHqL1s>fY :*a5q֩/=Bgg1`O_K?vED8  i<ޔ `TZ '"R(,'1cJ+kC]"gL 8)fmD.8XQ472aI ($B py?K4Ԟpq,OkմdW\q8}=0r%,)DO$i0xOn"\萳:E&XS\| \ VӎPxޣ lS|W#\ T!$L m,3\GC@k$Ŝ=)E&K/teRe&$˭g YJY`gXD!R`21i#0*6XW"?K= 29[6ڙcI#ЖqRTrz L$L0cJHO9++Y5q֮30e{?-m5hk$.h"c Lqe(s kQi1qH&P~=[ Ml 8f#<rqƬa.9Ŀ$M1jhUlZըV~'eX38t^8:"&FC)!.iCB{ŝ%ޓ9>G󾄹UӉWk-~oL/4%NE zDS>'"0M\!.gMN}5- Pb24:g1@˒#C F(Z!ڠj0z"cSMCvwA/eb n9xNiSe DN/K7ƞDD,` 6QܚH bWCXA,RZ ~}Z"~cA%`5^]gxT D/mmypKӅU[tC\54^v$im q>Adf{(en+]VJcIȃr"{&J9ź P(cK C/DDޡC/J:&!GCEbTjlYy{}/6hzjð\QXmΙڂ$%z>` Q C6GnAƗR+IDߦOWR N%s@0 -@5ZJ.ñ?:_mnF/{U#U8K([«Ę"k ␔8(k\eK@4}ST?$ά SOHt6 BytW+72Pzp1Ϣ8f)І:3_ʋxRT#fR"$)!T ׿kFjѵq̔:R֧tgV<ߊ_gO~,^W[g1D7HllYAN/2k;X n?![B{Zնb{K7uͰf4cͬNK\!̣w7ɲMos!k[edwNֵZ<Nrܰo'R'_u gtXԘJUw|ot/|~||?Ƿ/?R?i9k"z ^ÛVy\Md. &\O]Y>A󋽵\@r ЛD}7(/rלqD|{m_Aqz(ǦTwT&*D#x;~yB"vHha9971[S0%픕fe8w7;P^ucMS~QmEx #њ [aNGߣ6Q1k|M, *2$% F&10_/FKQwӑ~>èƅWyW$ýb8yAlTz&80\,+I-iDK hlN|#zN[0`,)eV2໛Br_7jp a+@*պ Ȭ @8e-Vf++;g&橡; $caVP|)6L %DqջUME—-m-7JV}c 1Ahnݡ/Aƛ2C4 \qι YWd&UIk%gpΎVc<)q4xlTht&rR%y(rdR &N 949(mDVHu mm"aؘ8h 3E>ÙG>6SH)`@ CZuSR9"GvaHo41NSkH(DqŜiRƁ>PBZA=#`A f,@)qɹ[h3'<*$ZYR0n5׍!{OxN 9'B[ǤCvJʩC#K"(AU ; X ) V MQ-XrNg_Pǜf*GLdm2,}:#_=^W134mʦ R/һgT:덵k8dZ&gXk8 -hk}0=|@Tb?no4'qD`%,bBY0yY _4\r<<<-ZM˽2V1'TK-c9Sih'AFb"p!(Czn,$&. L$hd9ӸrC瘒> hLpjS }W]b3s4ݍ2v;>'ov&Bg>J:y2R[?->5H]D Ǚ-$%LoшЂ6y4Zd2zd*tуJ,ehZ։V;ճ)d?2D vWk'()c8,]`t1+zdO2~ͲB,Y`ɢWkxռ,WJKѿ>2燃tXP|=WTxneOD΋95x)/b|s秠iqrkmg͕~j^ʖ苖 bGTx\I8RRJ+9_=f~Ors O& ($(ML@9D<a c\lRQ)IPnJtD!wD1_h,AFO2X O> ]Q%6(z MTf۪(^Ḑ'|ƽRnVX*renQ3@ŷ/ ˄ãe d2&UTP6Ky;QY&NSǞzrNb7eP{;0x]h/3i,5 ȍ0=%7s=93oy4t+Pvp SL `I@̠\=V>yL𣼕/޿M ش6!WӮEX0yC6ׅyDh7eT+A7Wg`?Rt-Rt+ z.ӛH}^6{Sr~$ThPy$;5iZ Jt<!AR$^ #HƷ9cj0j>תJt&yp2Q@Øe%F0q6=:g;&έ6Q']|˽\^-(TECm$k_qtڧ.Q+G&gc\db݌y}wރ,|[6/y,ڟ [5X´oGNW7޺? lr-g@v /dH_>;_,?\\<<wnͧ闛-YMJ >Zgϫ+h/%lm77`})8tp6ݗJmlέtTDŎNF(Tj.t L6_0bk}heDž5EUТh$L' μ/8s_WOrR~g2JPBb*XarR9d.*p&BU^8%p@zxhv|h/h) CI#= Tμ J tSg;5ϯJ/Y)i'ogԀ!˂3BV#qIrDo|nK$6sss6zR͢T[gwzsO. 8똌| Xjt]/`qgUw{%/'~2=B1PJ pցk\:݉NxJ%K <5GcmAsuZȓQeJ,]OB)A:D[6*9H";2:|M %6lDEɖJ ntiQeۉzֶ0^w}s{ J,l$xz >n1檇ߓtoGBm٣I7ܦFwQjbE kx5{@֖pg9.u=ƕl:1=#W5/VaZ]~u5ݸEcחZf{C`ϿhfWЬMm_6Ga Uo;f/čo6}מ&%OXct_*[SEȠRj1v@ s,tc'u,sY~v\ģ ySϠ#&dŧ8@RAW7\EL'5yeqؔɣϨbgxXQ`KE-+[+ΐΘcR@19J.R ŌpDhky  ? 0&'F,4Ȼ<>'HsÇCNm EQ!F6̅\MpObvZ &6>{SPh&lV;Y`V5x*۲(!bkLgwy3"ϻC+ +oj(HV U*mB-b y]OyգBDc;|exZoLP\6zl&+Ud"1ZMϲC(BYR-:{(A'ԳHQo H<\N|7^]*L \ %) )'[MNڑ. :#?Ʒ5O$#8R5XvAUM-d'e#FEA^ӷa̧*g>})-.fNJ.Pnz{"??Mߦ\.IΦ/Zy9i1\Ngl~)kNQ1 AV%H~Lÿ:Thw}CJu9?ql@oGUU_Dz}ļny?*pēj,xC뻏Ӈ< .?]M7rg|Y#}) `~o9~Tj >OD+}p OavF?.|OeD 2v񠱗)n { ZRWidכD}[Ǝ: =+A)<8]2WJGŤtWf& yT_oۉtngTi[uuEZ%ڭö<HLÓV|LVI,x0=XOT煽 I_J\tMQ*frPC}UFJƖe}3/ڦӑScxN|}s}#RX!gN$ݎל1d1dok.Ul:Z#usD&6nDr{>.(=9qt ~:"RQCaRU%ؒ4@~жT91|r!v_cɱަ؃˿M&7փrs0dLKv>5˒[5, fpqOrUQq21Pqd]u}\hB8<7[.~hg[l>?zzl~zD& H%HE\r ̇y@jWLI[(h3@mm}HG@*z5VќOsd9AsRFym5srjM8l0Ps5Ġ Munlһ9>e ?GGj wm.bl :EQfm*jOڋԘhdH$'YQ_Q1Lhik9~t['o={~ zMTMRZ5Q&%}ջ^UoJWd*s%RZmzJƌ.pd8Z!"ۤk&FnG~\v[b~“'ޘ|_^0der={s3k[vVKlJ$oIN&8oEBeK3e5 6y`"*X \;#v7q#6Mox<:EÈ#i)~.1VUFjX=9%H^iΰ9+}pA,daĊζ.*{Sy!QF"U:a7q¨_.5`<83"∈#",&k)b,ikuL,;mUU85\A ŀ Esצ6.m9 g)c;G:؂9%mh-3"vg?"~XdIpq.pNnX\q18fl5v b"hEDb&ꨍϵQ4YF\<. v[GU}p5޳вzb |瑱 1=qvqu=Ntv=Q%C 4DN\-!Qc ,ƳY'(Y;b% 8WݐKŀ؉Blv6Dc<ʑ[&Yc;OHhZu"1$pB_Nx{[{^kT-H5g۸p)(ܚ ,d滪Z4-ϙn,Sy6IW_WUQznσҁrA+jHDCT& Is;Ch o'oʞ7}S^*}.G1ʪ-*hVj9Y젆;,+3-op070_ cdy<kKS@mV˽ﻉs@``I可XQ^*ol^AnG#o|UY.[~dĂnqÖfq܋ڦx?A).=cT`Pk%)9Gll*몲c1AXͬU%zw<٥bD& Ɣ*8O(IyYGI!;˜cj-^T̶YR5`Q?{OƑ_&x hZgM>k61Kkm D• SGg`-"*mdLe+HdODf 5ӿ)1B#Bv]VbhU3Z3¸I`.{:KL#1h< t96+|Ox'8Z[ݢ5er$462tMIiTޠWh֛t\|"Ztm2LuiDޚ3mO|lu8&+ͳTΧ1% F a%P.X!A%U۪;Qm;L= dTkpH['(~=?jo.E\vLj+?9i #o`+i\ ,M4 I 3N\H̨ms92Ğ2_l{s <* dHWr#9%8ʱWrzvO{"(hLBx7l=75KNe` ]* `Z.7҂aa9+WӟMnc:N\#HfD+{04{_#ώb.ե'0J{ݰS4WM͢DnS?&+>҃3SiU4{>^Œ <|20b8 ejj~0][av)PDa2ndȟ6+GꑎW +A,[bĘFF U,|zt6 # rT`_u>UkŨr|6y |6ɈWJD|j0=_{IZj_/;OoW_~/_*>?_^>÷ 8(q -4w/~Xhƛ-P16g=[/+vphs+}sy%iڣЭ  ay?<Rt M39=tv*οB`7r W4P"T4pdCX~Uvn%G:oMܬ(ߠES 犳Ew'l]^UcdYzpt*a"U! oWn:`yp6ḏh"`yc1(t.9u-<&H|lE0Pd#Ȳ~+( s \K+ti[FUɒMV6*_5Jos.LdY]avVoIȨH+j6J!U\[XKT]Y^K%Q%#8t%*H1[Rrn3F6Yu3n> YٙAٟSCȥUdEUNF+KH]JYδΖ{"I2w< +x[0dȨURGv;۶YME{K(lF+sρ彇Bg-ﵒx8Vq1t2):z 0~ɕAU )HϼLDB8G3]+J ?W39ɔūD~,Q7gQp{,ࣶ*ũ%(}5/6er1~lC>~h-ew70yYYgK6"T/x#\ e98o6cw/)Njy+;?fjt5Ȝ*0yHO^Q;q\|yO{F`ǵ >,pzt(tBFHWcn!u)SQVs[Bn]dE5V1B 1h0fOR)L*x2b6чÐHܷZ ^D#U) >X[gV12:m98kn 5&O)~rm:4^GiZk7>3)9,W_G-lcH`Ir+SNTS:bZ`:Yε;U'37@-1U:#s=_ig}|< yUޣ寧^Wn.@WTκΔ,,I) ` }*β)2(_i~L|Gv~+$RѥVq@L'HfjsӛAnW J{1Ef +T>R 58 ߍ?uv*T EN$}rl{WqgyRtE^,˳y2[pF6̤ټo49e;?5h5[MwƓa44@7TTO*ثً{Qcu: ]騜FY[K"}Mar鶅&ISqPm+;&󶊷O'%ٲnK!pՒøV+҈.$/Rmvq/묶HCL?RFhl~iNA_kwK\\`V\ ZaF\ J% ;qu[sD?%3W#׏~y'\aZ{kjB18觻o4'g"7hjAs`V :qQa5 appV"[5?QC8d.9PZ:&\rc.N{|dVG3Yd->ì]Co~w?E`|`gTK);JrNi\dV3 KsI'nkgjEڳ4|%0"ߞCX6YufozkCpu{T ,R_?yƎ.od8s'GbZ84:x)-@^I9xhdB2HL\h/,Mɢ"RVKC xׇ%Ւ-LwX̶qXzED3Ϫǣto;28/E?zͭz%v˯m Y_m6|IumQ YCS6h.hv[X )eYi}fr}|gK9r"6$h.ԡ7uYӞBF'0FA0MWCɌV`ѐ3J1BB0Rõݔ"U%!|ct#^E^2;ɗ2$1J$&]TY )p xGBPXGaL."fT@'E[  g"L&St7mwT̎nH}sOPj'Y/km#GJ#8`g2ߏ!d/f3vO[cFx_Z$ ",~U,~ 6Dž J)|j!|k.9I0͆Q`4hXT(-ȸ䃙d^u̞p..EcfL&$ݚbH)a0z:#ԹRL]L&o{`z:U`fE:5iatHP#*) q- HםWLih,>GC(J‡CNm62"F Ft|/_L9zL$)yVU տkZ1k53(U^>Uxr3 wwŃW0j1ksw{݋rl^풣/`>08emn iYRʏ~]3EHm3˓ a`(|8zhNt4dmmouɺVkҼIJfk$5am8._O@,yOpg2 =*-=5 uuG0߾~myyw9&__{_``FM"|L 7-QSMC{b{4I|k1i58[{k(xN\"f?ikUM++ 6,UURm]* ;\* Q!}rJViMlw %'ӝXJmo>(ywYYT`#)E|__#:hByz4xn=AN yǕ19Qޑ~>kB0W:I`~42Hϳ/uI&8xӝS!;_%=ln 6Um r87{c,f0[-3ڹ9БP5]Peb\Ͻ 27GTRfrP8g7˲,+")Է֪ ,P5,g=,ie9 ^HÚDSҥZ9XG vF"hPF͝QFG YT 5CM_nL+x# x8ƺ#/v}={ e) r]ZdYJ͖r|7zyW|g%<-[3&Iu413;Ϥoo z(ז0x`y>t<@.4&ŒUXPBa9$$V8m0`sImc}J! E N&`ɦ $gaMan>!oXy2: VP6Eur!f`Y160w!HRh(5؂TD%28(O5vfeNPFk#;VѸ jU5S*$6xјI7R>x]}o}!t&MψAk]ZZ[3F8v +`vL$ kvtmͣo7;Zw35@oa4y'- e7%7 EN93X?OAmp!boSފ'ŌNj#rӡAA,u0oH4FBC!f]sL6j(%rKdXDc܁Xʋ"IĬ%RHD#iИ8Ժ ^kfh+%%w}đ ۺ?}p͛` } (\K\YzEVZ/`.pgrSFx[ϒxn$")? x;< 2JFN!PB+pх{v!<4}Z-Ai ,\2?dh,|5 rJ(ѬRGŎ~Z\4ʛ/'?/oV1#̞f1|gz=Lʅ^x%cadPvtnގ?d`]”:o,A/(}oN_'FKdss=7%5@̆vf  `hn)1H-h[JjKD"8"&J",A 1,1ʼnP:xCR lԀE*ɪcOP"RR%nbi1g;Jaظ m@Nc7sJ8UgfPr^U߭nSr'kX aoW1eFh˕3Ek ƈ[!\{],J.V)-jҎ.q,*ؕ[mxE(8Oy^BdH9ا7nXLes*r F`]G.|#pۃ%G a'P)SaR7IJ;;&%_H rD G&Gf~q {zMp*# $g\Kr1Ĥ-;lÛA䄜UnQ B.Rt]_4HEB 9#V +iiRyL @RXμDv fEV]%: |G#'m&tss0AU E<󸧽ou/ fԎ:F|}9~7{m]G:{N*/`k=Ӳ EP2y5[vީyQxwUt5/>R fO-Ϝt?.EtWwn=;g[SsY}yj*[S>ER; 3,6&\0S|iϐy r[!@a)[K1qͿB[1h3A`X( H& gfp:D8I`"@B(_fl&d)Blt8 q]̮w4XТf? ^iˮkw 8 _ȆRJ<-*Aw .kfTp 3 2i lJʩjxOܶdqb=@m5f(nJc*`*,ҌzŹLvݮm|g 0h‘)Eq"b1;9GX q!MxRfʁK;q?''0E.+Q1 { bzK;<uKrMN%,tiTJdtuQɼ$p\Ffx0\}|[읁DM'f SHDg.^;7/Zp ]xOBQcOn Lޯs\>s5ر?:%CW_A?@ɪr.0;7^*%{]?Au!4Ev͚/Ɩx-O\ugٿeP"arsƭˍ$#v?_¹$l0i˦Hb;AUϐ%QMq(SLW51熱?_~χg/y}I{<_5guͩ~lgey ~?_}3_N?ߟ]//˷}^̷Q"5hµ&Y{b<=;b~vt79fܐO(Z9Adq"h5K;m7i~y_ BCʒ}yJԞ͙@:Β/p%z^OnFq93)g%Yx>\ܤ!rc+g/նgy!q$#fFr2Gtq.E2_q Cd8&nP_Miů;twLB]Nqgީ_2"j6M=?HfiI[J]L,;-8"E*k*QDF+QDZ̨v"dfA=zWE 5گ_o+ӝʫ$^L^wTv^Wu0ܻuنLǑthD=O^9m+6S"UĺI\5;I2կNi۠mPAv}W[.:NO+f##!-8 l,zhe]dt )Nh9@`b)*ƣb1hNrٹ\h72Z8.Ე}j̵֨:m:Rpe|@T2(L㴐" Nԫą^)!B9P* ZB 6$o19XHQH4G:^kjBeAFvAmsKɴɽs C?̪2ͷNXmU=70I2\+JuE#^g>Tr\SUͨ1fix+cbI#uR49 1Y"y]3*ta5WʺPiWE˂'wl68 lc!Hdhr&d6DMO;+(' ~shuFethcIS7tYQBeS1\Tx/jS(xQsuFێ9V2 TٮxWLhZܱ֕v`7lT(G&iT&#JM:)8H1T . BFd F]  #(<|2ҀFuUևٮ~TcglZF455b7Vc^BW@NRg=JiDf Y;栭0 :Եa* h' ٬4QE)Q"ВPޓ0%FFvqlYKj\^b7D0H>ƣ(G3YUs"4)@тt^<^>{CA|wևPa[* Փ=r7$^w96gaMAp}qUգ3DϚgQ'X# [TNcāuF3I6$w<@v㪗{؁~3Pr׊K (:$D 9Wst 62S^;Yb[(Fk*3υ$WjTqY-*kZ`]Or~1Yb WOoNA[O$3T[-^fz< ǫ«K~W%_|gjZHa< 3.?x|~R)jS|2RVߥ6pp_Q}_b/~Z_ߋZ Qٵ( e \eqcBY*r7ȹ_ 'g->|M3(*@\$t1&p\?u$ijYLsש5aY6ͯjnve1A-Zf~9Z{<-{9m- 2m6JWngOȍ7ɟ?zNf#qcy٣6w C! ^p%+cosߧ|~>=z;oXreŭ.U dN˚+9S4[0;SD 1P4FXcE8-!<],žouUX^ҼR׵׷|挿]>kU[, JH)P\e21h%S.)!&˨<Υ*uwal>Eطzɚ,!WnV/VQbo 1Grx}" -  d+9N% ޓ' h$VK d\nky)-]N8^%Ļg}O0R|?zy$):ļ}|:Ѝ`V61o $d p ^dMzz*y'X=}DM d28嬔\a| s<(ED*LAfHY0crYDmp;T.S{g׏nA5`wp<eﻹ}7O/߾nGc2kˉsj@Bls u©5`ӸzvћaoShCuY>]5t67[DDn@N%ƽ2ށ =uCsy߫_( naӴN}]z3qy ^<|\f'gȽӋI hc:'(D-\0J5OQ*z5ٹtxsi\jM#0H6aƣ77}táX1m"}@OnykO6nlpE+sud?L/Pxî% Ʒ)\v1ZȚk77!Rׅ#t/>K4X=aoyN!ݚ<_zNgk}FEN$KK7Fxi')?pHB(\^ZMb{ !-R٘J6=J/p6X S󪡃f0{ʡQBd4ǔ\NO (ɣ.E*V#gowQ<|\Q !]**ԅ5R/q迹<ԻdGUԊAHg(7&,x) Tʜ*+T $^UkRB B -q 児Iĥhi-Xhx5rvt_} AzI鐤m-̣4gYRru ,-H; 8!X Y 9(lHG_P8S^^ƺ$cŧ[v/:caC[Ź֏[Hyz.K|2/f/# CL7Fp<'4'->D>[CUT]R=XJDVm^f!,ofM)EȍycenwM>Jm-ԥ[fZ{Ko̐3'M֧k~-7|~q"-k^H]a$% D͈G(-c}$jʑܱhA)7&`hҐ,*c#֑2@IȌ*)P )ELU3Ejtf)pVz=W`5rvt-No@}ĵ 6e.^Yユ#$jpCP^Hyّ w#!)2y,5؁WoQ]𩥀C]`+&Ѡ8r*DA$! p)$Q"SxƓy8@O/C@A+Gk0(tǑ)H$FrB9⏎JHhAL׭f?Q0r 8N>2r|ZXRgurT,DhDeq!1-z,.>kx1g.:v'$6w5=Anjh_{J_ܹaf~^zvWޚݎ?m\hN^3:D[qOy傇·x˶=uD|>Փ'{9le-Ւ,1l0<6dG{_dǞDsy <6)bǠywi^"dm|!(k pK=1Fv8?3q\+QhxmSɉ6IӘS~n z$&S+] sy_RZjqcHS;'ܶ˟e_/0zIwg19.Cj|vISCV^QeǸ)պ|\xIkorhHڻ2Ϯ'D4~c7n&ucMꗫ̷EWækBLv>(\r3^o{݅(PGڃHݺ&0Fֲq;~Ӆ\NWbow᎛E/«{:hx #,oׄͣ;beCnØ-,;eWVs#J, 6  F2C &˅EK9z4N1HODґP0'Ѡ)Evq{j>?wB>ξ}ޏhpp>$=ZʝS2ƨ쥯@{œ&G%r8IWtVOl~H+\L6+וW#mo^\ClӼ{UGHޟG膡 Np* ̸be0. ošMfRtB$/+uV LQTZxbj'y6ᐿ1nޓ`^zp^&C2 (ҮR6*Jx4rgu:%dsP%1y nڹ-ןcu 04(b5Y To㨚uy|sB a:Nn{svl^oF= ,.?^{;ߜEVk}+L3>@ y#|843/(q)+ 5nk(3}$'4;.`CK--iE`ZOPvd)Ӕ"IQmijef(5gSaCv{4FeAQM)D⽕L\4xM[Jk>@m\~+\)k26!kA*J&X%c>/q*2SKP̔U(6B HGwI-GPI4v NkkzEm#Izp>lHO}CxݙruW^Ӫ_XxV^: iPP!Ptb֢Sr#z"BHΑB!D8l :"RR* 󥑱Xs#c9R IƶX c:`Νyf=; +MMl87t#R )(A0';q Y:c2X$uGƞ pg6^ Dж#`& x KMjzvQX11KIǶ >k,Q,$|$% j:xkU2!Y@>G-~8"\J[-ddTyԈ5!XVTHu!@<Q[[5g=6F¡6c[DˆD< 9B9 E3ex" 9Ek#e ЀՑDEY%DiqS'G~T~l.SD'x[7PΙګkA)06xTкDwh-])Z?[#(iŤNPPwG&NDŽP 1DJ\{,N@!o[p4Rg5c\k\VHT^3y[P<&ХI٣)Ŗkg[yp!(!odZ 7ۧhg LʳJb|#U; {6.2/5+F:m0`)M"5UuaSL}`U0uAZ꓏.^BX61k9 4q`;c"x'PHGu:p2a=bfpw1+ji},eK"2y|K?JlaN߾E9ʅBIw#U\4zP5hY;h<6ȫVW> cQtkaVSg(sH_߸ם.Daq:؈PxE o}c5ߎF&\٠<oo ! idI6+:Yp֝nޠjrW^_5њQF8iƀK ksšF޷N!=oV۪GWO1xLjTj`K,`1+DOo'¾!0 hU׈3l򮒄\L>Պ?痋aEc +ݫj MHRBo" VZ!ONn%8uRRme"P Ad")!Ȩ9eǬ{ܦ6z>ϣ߬Mb5MV'x!Q+c>@ Ry\DpLܜ 6ln>S^/Sn"g }H1s][A]›r ZK[5{'ǃ,w#91Kv/LT.utN91IJF|"|9qJ\1ǻw`b,YA5d+ClMIOE;p/3V,xWj^_X&hJĵ#䤑hYKP#w JHD죥FwE<B&BĆm*QfE G@:ueepW]'>v>;o٨Xy*>p!H28 Q$R@'y $WR''IEME5*Z֨h(UXj2,D#kC+9gZ3&*yR:XXsk<3x^_vnA5OpN'GO(+P)G"8]fQ]`Js*/mkCȀ%Ķ /-1isC9e}w^)v Κo{q1hZNlb8pTٓT7OmG NyRrT9+)`F.A2.젹4wkiANHiƢ'IHx$,QcA76QњY,׌3eEAD. KQkC8BKwq |ɢ|p& b~5wFr硑Woy8,p9[VMj1Nz6{׉DV WAWX$2;mLŢҶ5~P6j{Ecñ(gPh2qB;c*x? *XjCVe#k}tR'd @nzv>/v7]F*|묒!Q떌V]5VVYiFq-R$S+g牝2!BdJQ=S WSGjS7Aͱ0/[-uPz7-]nrjUN6Y禶<2rבPJrF()(҃9<&FM@w)wU Y8wfGH7yP"qHyVTuSU\/JL& /ĔFE)u mLڰjgXB)R)590D1pਵ)6eCفEh <L_ Oz96B)+'5Zh%C' -"ɧX HJ:>$[iv0]A3I9[sFD*=@=!FE(Vګc'P@ {Wt~2=*%eh2"$[f'byezS*r)g" H _ԮyR9I8=JrMC"DjI3E`J!~tRl+') Ni6!7QnHm /}*cwT٬LN[;.JXK #'p[-ʩ1ܶ#*B0/$?X vSSiHmf1Wot{O;~z<Jrr%l_vj$#hS5.Bsp*^; ,(,`OW?يUǚ&!J ExKq!ސ߁!ժ&٫&ExYnVڗ|Z09.w;uO~+&3$+.߮ۮqE3GSi fAt\?</UvS;xӚ_]@NrȽ\͈_/]xv~fmp>1^-= 7Kjۭ"/g]o񆹳r- {$Ӯw$2G:FJ-,/t6 l\lxnŘGgLwT%u=ɾQ{di3(ki_N=|U'mGhs=[i[;?[N^$I^dC͏zDG_RtEՁ!nR7y+/%_#]F7kM"К*fkpjzRD:C1\y\o@}{1JQ?s0 ,rdMr^}ݻY?7?~V)BTJ,Ȯ(Vg,VeE )'!PR6>u{wdm1MW^' ߺ /SN-mRFTE$mueSMǗPWʣg|WQ>ʤ[ oiX׷lU}ͪNߦ5Nxqjuf>((, .yZvpVӰ, .PZLTw;30SS/}q% U|_O^ُ9lhppE!p=c\k̓ZæwZ%6iO'2]V [6zBlIuJ2D?p1tHWm}l"P7zkPp"z2|-(,[lL1sEMI<41+UL{حvXٴs Dhz6?խ2-5g𻬝[FVF]Խ+wNփ?c=Z!VXwk^%K lYrlBkGWW9ZG6Q6_D}E/ Ox}Ou[Q/M\zz(iXG[ffބR#קKP}^i Y^. NLڕ_9DݺVM|:f ZxԀ~nxH~vÆ:HgnV'qɫ톥j6|iߟ~PF\ߌxTaQ[p7zv?MҬUd:÷eeMMMs+}LfP~x{;Y1kFg{nnO,3k ϗyw3F4؍o{ !)wMHFk{:0_y$֬̓ikvl2p_K*ztк߻zx%ݱpOJ:Z2;- x=LQ^ 3[z~  g oț{sB//nȮ̍E\9ZebȨLTA$&\bvmAޝHRFaVwv7%&=Ykcs2~zxkr)XULS&'FZ^r r"i$lr|yC$E죓` IkoSFb*eƍYRu"ېn;ӷWijLHٷ jJXRE@F>;aq~[T"J QԒg)ZWJ(Ŷ#K_J*K"ա 57ѶԂtSAnoΏdVcRJ[WR8ʪѱ-)BzߺغQ\>5Pf,cUJH30f;~5G :(¾.;> {$ZBɓ9ח;%LYf>*!dafT],4 *(cRh)!Y ZetȵWYٚ$JpM<[H!  {.`Fnirv: \`!IvƺW+c0c$l٪D.kpKlZ%V*Z3#aD+IAfڳU/cPB&v$:t(yL{4 /%d'P4.ޤ:ǯx4_,HM"QZ1ʃ$$ԣŤSL"2Dle U4,ftYH!+]Q9+ .D; H{(!xu(+)$2*KV˻eC,Ē-OŮYED' sVkAOUMIp.Mݨb@(V Cmn` #hbd4WΎTg11[`u8 mz)+0h@С $ qbrΡd ( dJARx.g7&|lR5Y ] * +#IPe<òAm+_r5q'I?&0VGpE @ Jѷ&u䫌4: p&JE $嵑e&f@57]\R` 1 D`($VH(,!@&:R_IJD~9Uƨ  h/Y;M\0%W@LB.%")Й!l]@(5x׃Ux&/ɶ`d؛2'`g4BQ\@zETQp4*gI8o){@_ jZs4 rJStѕp 5]FAS2ѷs'PZIl,z@Hdjqp+.ۄʥ86BhB?AG EA݊ #RdێY%礁Z`Dr*%!&^GQ0>Uɝ Gv%kOҥ`saRnQ LMyYOqhTbBGE\c= A&|Z&x^e&J4O%5,At}Q@VAkթ17Vx l >XDunT$ 8U(!;# );-&sJ$N<>ؽ_.LWyIO"qאE}m6F t)]Ajx1j !p[]%`\`#TM&#(ڽXR޹6m#kLIΆąP?&TrM\.(M<ʼn7 9YHbjbC nO:KHm-"ZGy@8HAtWB\KKX^}Fi1HSEerB"%GwI2#SP<zm8TW#g&XrRpA~>Q 0%bw7 Pp1!,"^ꆕcUns zD^B57^[ӮJt#,K+,3&Q(8i@dmrY%Ddz ysQ.µ4_nK0# c yD,"@mEXXPQM\uU ֬nG62$LXon{'1͢yz[B[aʪ1&{dlݽ! 70&2 vàD3`4Q` yAs ZJrF^xTz`@RYZqVrORp==JD}z)+/aO'ۅ b^D!2kL@9\vr㻔{]-tkgp)K;#0. 8*#ƋJ+ { jZqs_yxk;@?apifZT UY.-M \z1;\&Q )nӅ{R@o1Z%.˘| U"w"J"8JŞW QtIUo⺺^l^74I'3\T*gGf|W[\23RCRcB޹of2Q}fOu~ -1|<$SI@|f" 뗋Ѹ}km탹W;f!p ҊKR.}291F?|dsu3quqͤ-1W7ȇe ̕!su/~lftp!M:q]O Pv B{E=;>9vQwGYF~%WѿɠM1Z$Odu- r^!K?W/?KjV 1> 5[sǟ??|Zqv".^MW`8>j}ngav(Cz~`6U`Uk5UKJ]$=AW`bA2]]O=o>?3)%%o~7ɶ>xMv3-LJ WsVQ]PFVr .mY]Q)]-zeg_E]X\4eQ5g\r:/etk6 gYQH62cZOY|R{Lo>('*Ͻׁ3RqcVxCQU Z߭YE6PWnJK.^:ˌ`ENDNwhVlu\6M}4Nl&\lQJushΆpw5U6+A5mAʪȋv)Ue! bia c+ j4}͒v[3=֣VO5sL?\WTvVM[WvuP:~Q uY颩Ik WyE.S۞]um9?W$ԕ'2PyT:kK9_SHdfS!Jf(uzxR͛WޅԼUr>zO}Xu:_O*n]:`D/QD.}ʘΝ=u_5W*M$ru(],_{j0dfF>ZWʎNGJg:%T Kf:% p.0uGЛ8 ,^K]nOޟ\ꮮ(qznrew=5yy&8Ÿ2'W^qSkd(Ho!%st)TYz <~ O=[2=;MYO01r)k:]XXKH,p̍,R>'uT,ʲBx`byn8-\e8r ٓԇO+t \Ν}Uw$DȌ܌7;ؕ彶R)dsma"0‹2B?7I4J n3Gڍ|Tu>s^dž˼lR=),VSvՀ׃z6'2>>>Y}>Hg7Srqz8TP *,-Z4A/Z[%hqtǾ>؅NmYƦκ߱az-ݾ᤾~PZv4 y s@Fwf'ayzR oyޮ@[栺9ua*O\lQ`4 0إcSx~7=.'7( =cR^ LH3%ϸLJsN_-gɁKOFqe}=^5hm.?NOΗ/>g7tSɹ:6?9%^B6i~>})p':ej.Գq1//ECg+h 2^&4[/ӻ;,[1[MsҾjv0U Cqq;6}nMz3\<ŝǜ i hSchc;^wƝ1˻5x^̛ep9]f/32ww]K5ÚZ /b6+UKFy%b"q73>HS,ݴ$X65=J5t'RR]/ili9ڀ7577f~̖Փϛϛog3nsm+_]vN?1.b3Z?*1U-r̐-'eJac;?gu0b{~ҕLgwۣboWEDוL87׼UMc@uUōu5U$DmpmVh5:x6-F!Ho+#&mEzie QXRU9:<ҁ{ŝ?;әΝ=s h2Z|yKVUY 4: t>E*cJU**M06U,׶3E ;b4rzGۺro]oW0@"]\Qh`v#K>I}Շ-k2mwZrgϐgʗw~9b߄wM wW21JT:mi:CsSh._ x7f݀J娴k9ߡsR=oWQ&,JMI%mBq t [lKw&oTm8nئ((u)Z;*Xvl$Cl:gBsY-+3vB̻&K{*Z\} zROsi0Qh0L-` ֿM>աFGOVd.SN!7l-z7eIiqAbAvǻǤGH^<+3b:rW*} !U][HC'zJe-ZaZ2Wi3CU%`hT֪qx“zhmR-wM!QJ Q/i1& Cb1z-T粺8gs<[-td{$.{{3NMF&oc@(b=h!od:'*d$!mR(3U 5)f`nIg{[ 0yyL*XAr(n`-XPsw<46#j%]X'}Az!z ^YQUʷDnB֨0ƙM "f=>g֋T31ʌ, ד"D }BYyq9%Fڞ8{zX/BS>x*uvVqNJ%A8N;~At{Ψ9ˤDГ,Y((oN BKC݄ $pyL (nS Fm؎B1e-TĹc/;kiǶ^Vڶ=1fcxffj?&2K6eHq@sGNemt_(B&dEQ&KF݊#(:f,RPa5qnmPL-qgbZGz{9Be~A1O*- Qf pRm:p@M3Wucz1cI+]6KJgep+HPeX*{j۝>_. \g5-/~~7ҺdBXȊB1Jb&rDž9a2!"HŇiǖPpq;pa n?eMIp}e?ƭD-3P% a,o`NHn~‹Rn\Ot䭟?t2K` Ә_K$O\!IJ܋q?)M3V2>H H}mfߋ\$ҁvO(0m"wo]"B'v_;_杼?.Pй4Ra4zseW-ori:e<;al2ΝKj`w9%٤%Udr{Jayo@88MVNƿ_ P^P$Y=YEnCIt:@†hrjơ,d20ӳ =xlعnwҳ~z60qZd6&PZ'd:A U*gtXJE֍9!'(Ԍe-9y m30eEi 3\;ޯ%&UZK;[.NN!Ye/y=}ӛ; s `n|eӇO^K&H?QFUؐJY FYY# 4kSh|(=#igO Q1%瑩K{eHV -0@Q b>@Y$HtW@%Q3_Q:\h/:\3H뉐iёI @!hupk-uR*hb |C L+J.zuqm\ڽX;ne`- d<~dڣ6vT&`Kcr{ c櫵l;$Zi5 ?ĺ2,{tDH* >ɕG4:, 7\+ѧe@3i T} xᙑpzxm]mv9rىԌ( 1보YV(8( $$4 !G菲j%UP"H4FIJXJYrH06 gW"y2x!$؜NƟw 8ރF{;2[zhElh~FEꣾ(mBD3f$eB Y&ŝJmKHG.Ǫי&Lr<2<$r,C *GV{l5qnkqW`Ǔx;s־T \S7t}dDmϻef?~ߘcN ,Ch A!;f|Ȉݎ޹jE-LlS y\Dd\3NZ@R6F/UFA#qJ9`9@+F*SDQ&IBsBXa6J7VrE80!DL%}_3Iށ}W|5IZXOmnǾTtf]zo;^^2(nvcJ( ^s@9 2Hg&ufb$z˭gb!`Ȳ&8xfJ#\tP6TR)e0 T3ϲVHaȉ*uLK A,$NVjlhg]+0eN?-m4g$heN c]I1J'mYv"ٛ@MmCz?!48d+5/O Ε %i8@4 XA8qR'wа L@L4H XR1IHSx {- wU4NwwNl;vy&#"5CS j|CCqUi-qai(!Y7S,aU K%!tȣqbgP6)%\TQ}J\RY긒b <)u2v$Cǰ6Y}5q6m]jVp㵬џ{ #rDaT:]Ι)Iև%5S!kHAR`o9$%]MMK5+u@}( FaMj lK*S::E$O2V jB4Js㖣g2{c]3֍F6@dvR,/4BF@n""* XT9%k?OG61L!"du*F1NA -gMS 子ZF$TސhH}V{ksB l?Ҍ{(Z}pv ekZ= )V);e{Hei: 錜^fqn/";s&h!.XϩCOhQ͊a1FS;S1N}HD=je~//Ȧn4=ڔ{q{4 ]=$zFcQ;9R(aL4#NdWZtiMK) kR?'lJ+ʌ>|~ԹnjH'6h~zOHۃ߿?i,9!"|/_4o~ܼij5M-6hډOە{~Kq5$g_i,ϗˠOYt5,pd^x:[RTFT[!]ߗո,Dcu4q7^~ޫfvڛg\y\W7?@ et-> lޫ$%adN#:Z'9Y?ӣ? f}Ŭy6 F'#P,K(Qa}){vdwp*Mk󕳓 oY@cň2/B 3 I-:䃄~"N+WӉ$_yg[uK5Ϸz|+PڨgTT&qϥ:pv>ՁՁsVdm ٿh߿Y+xhh2M73aDqwqݗy P0I#dϖXKړɆܭNEyF[5VeiJVٻn$W}y8iXn@ff1>$ŋiY,;N#w]lS>ݰ-PWź\D'D O*VP+0:@`BZs J@"4 gVǘx6=,GrS=1uˤW%.*56>}s~,(d\J0>j4䤓҄l`^ M#)ym"Ke )۳^&&[B>9Ŝ29YLPn6Yعq ץY.aUzTC#=G=U^HgB$Gf -@ڋI1Em1T>=~>u.y=cY'g?F5OuB$CV ɃAhȗlJ%Y"dpFKֵ`[/@L,\%mإ}K7*Y;nzˠ\w=K'ƿݯh*-ttwB [d4=in_n'K .)薒*Kyy9iM N,МX{&۽_Aj-qH Җ%kx)B4h\jGW|_EquTRa{eS3*n8@]* Z蜧ؑAlP<*2ppQp@QhꋫtJ{$E-aq.l/6 MK5H**]ɀ4 =S@)C2y^$EI9;'<$ O67v2Ϧ@k37>Q1!KAT=Z8wXn1Fϐ=Vк8$ykXi.$$Y勒2] m?$`X (p,06&f\\ $ o ִgK^ u\9GաNy2&|JrFjg`Av(VѲ*H5]&R24dX`C9>TzY{> ;%0m]f24_h@[t;WK +x?402S MΩA' f&L1X ]&y]0@*^qd@^QD. QMo!9qƻ9ٖ|c<^ݑY%[m= QXޗ3OWNEbțz'0Y+@I{)}0Cr/5 RʩxTgcmH(h)P|: 5ڒ#2@sN8rlUB<[ғQXӘ6Ύ'glhk WN&7>o{}l:_=5!N7eŃӮv26.{y')]^y6Ə.vPk+R9zέ;nm-Zos2e}|Bg[vܲ[ݻ4wzje-zrz8N7p}uϻf*H薎^FGulKsu<6ϩ~gC(3$ |]st͝*{j*p_C:Z^iRɡgio)!>+TJa1fc$E >8Vnq>{j-$&XD@ :לF2En$rZ+ҷ>|mm!YT,+NUFeWm7wh@NQ$S.kU) Pj  DBҙmu- q``x3ٌƔapђv& ŊVژ&X#ġmz)jٝ<wd#ī//,dt{K up(1 BfgMpAj!|Z UL늬?sY:[>8~8IԟHPSA''_/ꂭx!uezZ<\tͥv̎N/3}7gӋOGaj~RIoCG3|qazvt2a1?o_߿+so ХhULk`1P5軅s,~צ< jGԡ/1bN)y)_:-r ןNfG7lG'qyE|j?oawӔ1E#ќj۳|ڄE-mmt H˙'g>{:im6O]YE4FN`[  sX go koQZ:iߏu>ͪ5=m/{X=h:./*',T]uG?.ە7[B|R\~zW .Ԓn6)p1Ť>A&V뜅g^ZJzUjCZyS c7-e$|W+Aޜw}6ZﻸbiYv:g-O7L˿BSw19{يHT;oUmPtfg'\;e=i[3A;Gl{IZ.b,CJ`>kɢfA=j>-)!j #%x!Bʲ#k{~XlVޙ"m"{*! &G}_SN oxD~29 jz`lH(Xa6{S]>ѨO^3" *c5c9bD[?hlm0ڋöDZX!G j 9)~t}+{ҐeRTh!5{˜P6J槷z[DwNb}w>dt:.K,F;("0l!)ɴa_u[ zL@pmn5٠tL8aJWfE>a,ӥ0+*h彳<:uӑAXe݇|u¯.ƩGyʣiCι]"?rOn~'>~ݣ_<''O5 U﮸XC,>q QN~]Z ^n\o(/9咾KeKZ]?Pz=g iA-¸?-Y,y7cl@2JY 76tɣ,?xQT Za$>^HHFZ)]1 XjzyoK,X,+t:Y _+mBG5*8mt42]JT.:f'A&3 2Fi0,E TM57$ K@1ֲJ9@1#Xq:ҐNdp6vAgmog?fH4/b.!bX-QyOJ6Y ZR(ʡII!KA5'1zM8[R*DZs( =9M+̸ R?526vdlUaaX,4 XwKwʌלh/.XcyhЅxi|2?;FKҒU(UDŽ@zyU&Ȟ2 u&XڵuQNU\@("PPaS d'[X@ћsѺuK[ g;btڭVcQ6Fm;`7lB Ed6gb9966DW"k %U5)h&QFTtlCh͆{~P[`tV㱈#q@č ٚE% (]S%Jœ@N%Zuv 2muz"Y+U %iUv1& XI$ g;"aPCu6[%E.nZDUt5TBG1*D&J/RLNDFI9(X .Gmp߃pְ8nr_@Jx8>3'z?:ivZ;:,)C)eЇΧ l%jw܃?w|X;:@ ɒKd h]D&Q*jeNROO[]$]=d-Vٻ6$ewUm8d9${XఆOK1E*9~ÇHh)R3隞﫮JVgA謏IE..>K3E-rm]KMjhvrX<'yy~1l ϳ O߽%ēM2r-=6U-uL, Z)6f3~@]|\)|hhHOͧhT^Kc3kvmJc)_f4ԓߑ]¶lc D>e61x> [F0dȨUyݶ:۲ `ov]~GGza]'b֞ˆ!-:2*ohzL664J5Жq*I c!6 U@: ڒ! mK;2n>1<;=bNq!ȔJ7ԙL彅9{g tH1Iפ.235eڬ4tƣ.%͇",Z;9vhOHfnz)L p@2MUGA=,+2;O(gFUY͕L T}@˩ SsPpj6. a%Gf& MGYr5ڐ$x#U) >&,KmjpC8FC2c5q6L3.4χ'M3tڇ@Zz!6trFUAꣾF=f e_i1GRiJh0$؟V&b:e49T`Druܑp ͞YR{{VM5x% x4^5ܓկp_ jN1]+:f'fIHdufIo,SplH J t7kepF3Q1q،cTj%acj7a$ +`YلQl"R9B4DXwyoFc)Dl5&! eH0%i#xq8$z@K4e| F,i7VݸGkWnbvz&ڐ5R3l +0rʤ8FEeuh-;1"^XBLqfipg,"FTڋ8:K#è<`]mU ҐA$gݪzXv€m8ʒ0,`22A&2)<Ì)> 欨gPZ+PY;*~ ]hpA#^RYc`,`KHAX62&β4NjCf?5=%5)1XּA. mY[B @KtF@gԪeIO+av`AzCh̃#hh<$ tI +VpT1UUژx`"EOvD>LlG15CS{p./h֛t\|*Α !n!ٮF~|lELՀjC'++lvDwdYI, Ls%n#BI`%P._!ڠתSz1cU_yvT/st^\}z,ײtGwE9!BJcY9gV{k+i_,ᨙ60y 4 6OP c45.eoYuFcD v.rdF+UhF$GPEpp_* ҠHD3.)FXƣDv2$1J \TY )p x72ǟ55mH("fT@'-BGÙ:?xJ΀0&y  !jp Jm^qHg9_ #%7=u֏W xGlk+7i%$?̻\Ta|B=zq2;2_lx\*z7dij#DsDz.{"8%Է'jwK%fW&0i8IR֊?Si>2ಈϹ~7.sAH{5$^}5k1?˗ɦ=1n飰U͢IF++NKq5jG4zK47]̛7Wma,S }:I {ݷ "j.lO`޹4x%[i%VtX mfq%Ei0lm`2ϳ_yv~vyÑ\*[=V>kKVeWmHiXpʭdrxď 'noٱrSk︩1ߓ &{┆ۓ^7'oN v,Vx78ⵟgˠOYt5'iSYr؋ؤ-[.+>azVߔJЌ<*ք au~jܗ,@;92Խ|Խ71[S0g7-p8~vv[sL)&)"YPXZgȍO/FWϛ3΃!KdE˓|ԿAs=wdfpC(Oe}GH+NJgL 7 ^pHSھW/&TW/~#oU8C?v ӐxӇXXĕNZw+,g;)BBtIzJ޶Ojg$$<9/^N^uCڠ":P7OtnN/]Zd:t[:#t !6Ls49~U;u&79=s[xE&e3Wxb|_R?}ߛa,2]{Fͫ`h4>z1~<+N/޲]YCWm>J}Lc#4^k4g`Dqݕo ĶN!`]6NT,aO:k2 k:E%9".Q%&ey, %AJ*Z9p*\$w*j&TOWqDvÏJ̍uj}cÇ?p&G#z(;mUkLy@;孢S>M/_ga{1x2>K(v |۹c =TwYJds9(uHT}!A#5J6JS: KWuG+,JGop jE!bLWRABJQbOKS6t0: $IlT0L,4ᮊ#خJu(~2.ɸSLd:R#1gXGf}q 3M(A傰.\͒,3MQ* }9d(ʽ ۭ8B7Kp[l{,}w7eY/λiP=taYAd1sYO5}7H6Mf34}_#2AڪܔFr7Z6h`ut7FAPa >v@$K}Jֈ 1IeElx $ ܚʮdP! ¡cmtշϋgmiݪ"5l_?(xٔE%E0ǫ->vO6!5x7٥PU0{0Ԍk99^ L9hDiג9>琢<*2ppQp@QPiiꫫGW@ T-ʌcf3! *flHkPXI3€xc֫CuH@͕7.gg4PTǭRwS :;$I*tY4ltښg2] ܁{r*,t]\Oq⋱:$E#w+mH_e^ݡ>o @v$>%ѧĘ"$eYkoě#cM[4{]U]G" *xJz51C R#?LfZ;'-ƒ>t2|_6J%lEg}4Ɯ6({"Xds/UMԒ>)gsOTfsܓw" DDbMDW@%GWJ!JHl]\ +UinfI,$yق!崋hT3{_pZ60 >wp T3ۇ[sʚ$U t"4FPSD$-L@NK9~˨Xlb?M Rעp&tT뇃pP|#F-R10P2YrJH 5|F[u1@QW/wH"Mks~&rnB TR!k9?mdD^N/Θ~p+`&spS\陑R0ky<˾y7:W˗!|tuޝ @5V+ADr: @I5.´z= Aen67 l7ΪpKne3$1П"٩`?>*KG*iAvjoE?Ϫk#i_{J ~f3@ X*̳a\Ͼ-/mF&:w%r,ίan4ZlwڽVN6f9fYd{؇=|DgHG%çUEm+|@W@0{XS*QթUR^ \Q$Г W@01p}R*Q+ȩUR^ \1(" D0%D.k2}l*Q)H W/8HYsD MDW@%AWIUM+QS61p*Q[%• X̞k.J j9:yJTR +E0'Ap JUudxp0Cd3.EͨUuu{[uJM͹bH\1.D%iӚH#\}\ i-eps0ǒcD a6طr-c1 Q\6&y>xLL8$V0#)6gF\TDc8_sc̱|:Q+<Z=c@93tny$XLjAʚȸnsYJZX@rFϵ,wJwQ%ڋŲZ<Ya4%͵XEaS3MR1@"ZoP) R5xBSw:^Ia{3hI3'ۏZL0[=ХD+W@SUWD-WJ-Zzp"E CoЛ\b:2IbKxX| '\.eÇ){^Q3%k*!2 RV ӪGR;8:tr;ڳ#8}+3Wq~G7?ՎXvN3ٶ˾-7ћ iɷfj $qeRr\a4 R^hcʰT0 KIzbU6zb 053@j7tfgQir<3çuQo_2B dDH38"C%$X*M/~d*~K*@\瑅HL>e'".a$PH8H\26rv@:p~-S*%+5]]7j֠,5kc|_Lyk{/MRKK?u^+&=b).Ag\J' sb 1NĬ6ܚE2/Wl|J [`x1@)¨efp\q j,Rg2DOL@P%S"!,s$fmlE9_ (/κ(V7Wug_g_@ e_~udo Q{eFO?M]p\ ~]_wGTx0w{쟦_aAʒ;r4 "kŤا0~xnwơM+sKSɎ{.mtv氊S`SKLytTN?|gu&,T8y\u:ebJ[8{3K mky܄{rNG3os-eW:Nfg0qnq,To,ɒVAؓOzre}p0,GEJ=^I)k>u:}*jCuW|Ћi +3X0}r~/9$]8}\T&S>[3{~[|^FϼjI7Ƈh^¥ SU*g"x?hJB$+T2u)@Z4Lݘ2u8}̇1fMMi9K ]xs>ο]\1_erPS'dr^pR,x+Mn'p嫿WTɯiJ01X/J^v6@QT/;4Lytq2=T $8r70AU=wzmުt6LJ6Qd$ /%];L:l4 PT5s3.7Nሑ4E-z]'az=:֪k;F{~9lAt1T 5˅'UK3Y[ӡdE}qX"y„/م= :~2J::qsS3~Z{9~̊!uf)dL{02뢛 ]NIr<6d4%ii_~$u4O{owG,^w [vME:n;g=/pCo*߁燃lϪ-{%C#_.:)Xs6Y߽͝ZNzU6}Wͱ6k:$ }Yskugͥ@~|z&GBijZ2K{gZHokz|,;X*r4a#A2?9HysQ (>c]"p7ĀUs}{j€ې edSh8|(2:Yv0 w떾\w`H}fϡPnlo;mroŽ~Q|2$FZqöI!h!uv)J]_/p3&Z&ĵ r B($@JIPpV{1gz4‚SѷV=C Ĭ)ןIO1埝dN]Cfzy~*]5mˢ'|+=G1^/E /n]Jwg5vh0,rk!M5S6'`(9?hy?YC#x%* ',FcbqqDi-8Jj$(t2~0ur6 X*UγFqTFRN!r(sGVBBPNlfLkݨ! b; jxT~8!,lIN{+%RZ97jW$u aI2jūmFU%G:,T!2)9nBT1=VFb }l ." ZQ(֢dJH1 N9ťoW12#H\y"&ndXRb E:f8N񚸾^ۉ''l~xM_;b;y0 kz¶4R ;ԓb :%б7*XCR'ZK(IItY$R:wmHgm,_C.3{ ng0b& P*Kj/bΓ\Yml+meh;u+{"ҍ2BbD ޓ-rRdr ,H}CVǖPpz{ak&CIU\|VnJ9to!'9>a OOA$IfC#O)ѫY^oI*"#U3:-4vR&d hՀZMbH }/Q_gw].3+Ŗ7w,#O0|׻eLw82/S2z.{"(/G1 tfٿFVVi`,yJ& h1'7~ 釓vj49W S }j7?/K ,̾6_+?̛jJUC@x|6: ]ECѬAц+%yexr ڕ]3MM./"kuru{tz"8Kera՞A>?̭s(6jV|u=KȞ+GB|Ua̤Ua[bĘ,X4lGc57u.QgjԕZ|QـM3)KZ>rɯ+x.&;wXknhMcr`rߞ?Ϸ/O'/|{W'o_$q ,$v ͇֬4lhap-|鄯W1PW 0׭- 痃CJrY(sY]ȯd>k78 Ml~>zazݦJFSc+B a*ek±D6M4uekaPgCF:JD,OL&RcƹRrg(=a&}9$(/S">v#]6C7^N꼇F}Õ{oR`:[b#RM D֟*bY:e9T`d49\Y#L28lde2(ሩr`V#~^>^\-'_wM*r-)&%Ve8%i-KҾX)Od$Scdec) @'΢:ΜP37 ȸd&r|!89JZhd#3eNfќD ^eF)DB2Id}22f4$$̦})o ABDԑ~i"Zp;=|۾oUSE)~UПiW!k)=9{D9DVE! &dflfqpg,'?Q*/iT`t&Ug]irU ҒA$gK|XvBeҗZ3L2A"21<3٤ T$2U#gC>;hY[ޚJ%+/\P"3+ >:L%t$ ,XkJcLe+HjCZz?G^)&ky9\1+gmh0/A FeX %K]%S=TfzN`1Y]m/\Ikq qxg;+Yݍ)!%!RdYJ_2l.0+De d([jރV \5rs7ػ>x3y9T9+<#8Rm>ٓ%}c2 &'/53de@PWq[F[*IB0pV~,'TV>4StPOEctU5fJiM+]kZ %<מKL1 AP5rA$e˵~eg0 s|vxw/>h 3ʇO$EƣVU*E%cI!H%&v[y)I3}֏"`{ܴ0x(Qf Go%Pd"XpdZ&r<ɅZ42PkLF֚cb߽F3?%6Y$74(Xlcd2eVgi,}uw tvis}w2tyӯv# ,g~mvzך z%f}:Mޘ+`I^i[9G߽KO?+} zԤDlX7lUN̮*BI?խs4_P @u巠Fo}kUig>( YRABJY?u2Ad4o'Ѡ"ds),5YbJC:C39 ̠-Q#U.*d`Zq)rF3S*`Df} $sT%Ή$tsCB~Kњx1~ +d iYO@]AGJkQ?YpBvئ홛37HF"}_#Jܔc t Jut?TBj@>qeLf9Qy̖BFA 6h]7Ƭ+x~^~gVw+*릲Z4/ ŚeM #k#r^$W^3IނI.Y IJ;q;ŚU5i)y1k ~>k5ukTM5T7[:lWDnWgtzd2S |XREޕ57r#ٍؗuĕ~=kG_vttlq4:D)%ER*E"H$Py}Df,455B Mdꖐ*n+d RRVnʶ`2x;Źw$}KbfyMx =6'gݢ.8ҕU0Q2SxZr2ZEE >^I65xO֙]JZf7I5[BZGב7k﨣xh i5H!6j7Bbz;AQ(ݲ%*C;,jIĔT\ LĬPTHX2hPZmYQ(y-ENZ|xGAIrFFF^ P[ dMg"~k`?; O ϹɦT &g,E;vգSY[ D$"(TmڕIFaL)M)aCTGuN%H &U6e09 eI$$oWlo 쑤#J֑z 2/iآ!+XOK? M|C:uVmƱhYx0DG$b#mx`<'s2 ~Qhj6퓌 )Ka lNU{C2A/c$c9}􋖵Wp_uV Á; jPɟJS򧒻P-K0 ľTJݗy%6ԕVXw0ꪒKPS+컺T+TWN r1*v]*zu ~`p0ꪒ塨J*} AkKp[gSƾ,'rNYi:-8dËy,@FW}78?ܚx5Gp1@ȷ {$@ӏ7:Jsu9-.8ӯlZ^3$pٮ28GI|?ݚA|6Ce*R S׍ĪWggj:8ci󡮩vG%%={UƱb;ãPTlDU*`Tj{T\L *'Bƒ꾫ΉǼӺHטS ߦF(o.>0KE;hhnKm_Vl[G ^T,)Amhe\N*Qc˨%H_蠥e>9_xuiGZ|yJiHLZ15xrF5c 4)bqN%Hc0*'Ee-IgaUʘA=1LҔ$%ҳDJ2%{VypW/+ 秧eݱ]fVc6va; 9d뛪 Nacm6ASѾv/S@# Q{^TEe2yR@ 0ICq1zK28V!ɰ`-:1Y_[S`1+)d $8Ț}6 MoƮ]ȹE:tNNN} 'ޣwzahM(>@0$)HhZ mFbB(1QPAMʠFLP^rf# 1掐(ؠ腭iuGn+,EQeOC(^9P`IHf,Q (0;GBJ/`Ύꐕ:PZY˨bT/k%'sѪ"HA!yfv:"Ջ@"y=z=c7ڲ7%랗gKAJzMRlRH&ѻZZr'=cAxx{`tEDl<0σWCbf#)dʲ#Z Nx+NE:{X2 ~vDZ3 3FW99LruE~x#f /h*2'uK"C9z<#Ig7]FDn/Fb Xf۴8T5b,6jgyK":s4:0>X!;`M/݉kٷ}8a:WհF?z_ķ4-'}O[Gd% kN#6aF%ڿܥ:Q>F?}}[M@YLxpΛu'fAv/=x8H3Y+ B %M$Y.kKN}ǚ:/ӏo-.>Q?".Gj&#j8JP 8xhkrp-<&wMU.|`.J{j\rs{ᇫĐ0 \{e4xmu`]jOǣosx2cr8)rTw:Qε|ԋz6dl}zzV`U5"F-=V/Ll8\1_~o~z??~? ? el֐@LߟDO ͇vy Ct99:%t95r{i0mk+@Jzqqx1i:9mwӵdTnq=i(˭2[,F[is2>[:e؟O, o';^H+lԚsYbo{|#|5|㘗x$NӀQ7g.:52h%RC&"7@{㜧iwdoJ>@SU m>4l)`*[2!4? -nRCӔp6ц 3;u o -~'Nmw1wboY?`^~`vajM?Qi5<5kq0Zs68MYb]$R)ɒ)Ml4gu.b&}7^VnVlKxQt .IȪlЅ$Iu41)͚sǣOö<'eeJ c3O¶ψͿ>z\j3Llpya Tr1x , A&5$luMOM ̋)^ݩ]2lOEUӔ;Yڸf(ԆmcՈ/sf# UkFB㵦FH"!=RRmem]AU|BjQYew [Cם;rľU{tAcs~r-킏aJ[/]>ZU0Q2SxZr2ZEE >^I65x78KkmHe/7~ 0In6 {0)ӦH.A!%DZ1:S]uq\J@M ϳ Uf20",pҀb6N .=Re,lB9rB[S]E`8ЈFU !hH>r "$n7:Ei*9cE=I&XsE;^@I 1jM8h-l W^Lb}4=w8apI% M ^RLhk5'Ȣ'ӞH VntG1S#K"gm=p(`o5Rb؜e *ݦDHP8Ng^Mǜf*ꄇzD3`^Έ\)j_'v1g_rq˯qL+ڪܜD3vɗ7mϒF YR˺vEJi\ю*ja9=a!la6>덦yoR[lhB @/٢?,k5W;XM~:3Y|ur3P$I僑8#F-reKỦy#ݘd,h'ADtGHQ* x=9@ ,3{8 n~ vn#}ɰ@ONY W/iQ5Ek:N#0v뻢$%#r%Ky 8C!yR)" IPD`C[GWըRqST1HL451s&[%Ҩk65ƨN*da 8[n&3C<jkW,z lI2kS[yd,ƽl_z86}⨭\+dfgl:^OQ SVغuz}8ɆN۬7=~4e-lݾu{ƛ;3FmRa<oqv=o3燃^Q "؍wOO/n" }mY g]p:.el,{noHD"Hk-Y| Tx \IPG))Ж/ԟ.<֙W:q|5; s!rU"X H1$IF95W3H:2oKi(ݖ(zvGjn~;J~LrM-Ah!<&MRP/1% QE!N!D{ ' kA5 =O(Q'uQh-ب}Ej,B19pup3o^ˏ Z+ͽ5؁-6(/W/ph <&*B VNEa:N EgtFpDip@zKlDN\E-F .WޕD7;9-x`p)~;8;͡&o7TDQwggl h;GA8_[%X2qiJ|]'׳qfp?c,y8=*S٦fW8:9gPl.:ogᇓ3;u695j|!(b+=|2$p|O,8Ŧ efV/hxmL=P6)se~mǓ_??oXrOu`RƧi|!C/:~lCF'*ޞ!nj.a} 8?QIcq6Яr jӳ߀$ZڷQ^*"Dǫ{0ǿxq|yø,n8;# %5VU1*^`96c`FBrŚ1_(W M=*[r-H|o8\]T,6R$Z<'կ[wٶ 1Y涳A~pMgfiދU.$qvp{p_}:Ô*J&JY/iVB7ќc6߬bP/WFk()"V W[.@ r|B=v^cNr񆴅+ufN`4 iV w,Q5$@i=u}qiўlhu-lo Y@3 GYpjŝ_'u9ߏ-)՛r NFƕsRr$q"LyQE@aHG%j@2>9~4 2O}Jʩ5PDpJ +j1q6+j<7ʂɡ/bPVITB JJj$x4!XO\ :u)|$Lg(2JR9W!qz^ q,tJ`UDKrK~̗FblFrJ1,,630 wq?exEN{rrO?OV~۠C7~).< j|D3[zzـiSy^}{CϞp{~#}BZ6cr)\ІH7  F8BdiץeI^1>6O?޻r /z$KKrPP`A+ء0veiaAd ,8 \eiwR. vpt * Ǻz(p"TtpJ(Ƶ< B˃,.C,}, \I%uSB`泴|׮Э]=ER }Av@pgWY\8*p\r;z:p5\e ?Bq WYZ`WYJ9O70,eU|1|w.rD۳ٽ y靖c#3X|XHѹvf_uzw6כY7ߟݙq_ۇWFĝ׫C; T=&ơ:Wßp*$dI8}˝O^ߝ_l j6Ὰ͗]s,6_ ewŲ9RJGf+g5eW \7Zt`'+djbW`l])f0"SוRzYt5C]s2G(;-h2.gMٕwz+RZt RuIW\ϝAPDF?y]ƸEW3Udq" Ab5nGW4u]vn,N-V&xfIod20anwoʵԾeJķR= H;O[B=UMTr pk`W,uՌVGre͗3^Nh_0? xpOjm2#BFI$t%[$"])pGW; ALfPEW3ԕ^*Sˮ])OSוRNmEWϢ+_+ťjJ+vRJ]PWM`}0'M lxRʩ'Yt%)R`_ϵ+{aO]WJd u}}u1V+~SOWJK2u])]t5G]@Ԥ+h"3AiOjeX QW10Sգ+ŭΠJEWsUOiC䕱gDlKh˭4l":ٔwv/NAˀ TQ`]5yr5WevyReFї7ޟqj[WpFFQ4ҙ-/t]=IwzDJ|Qn^BA7 CnNI_]iܢl^Iutd.' ~wt[ZoI&}W˓_JW7_-dw͈%{׻n`vjw rGqm ߟ{JRx˂{gx&NqGh9Rhn>n9'>QJygWw[/mπ F|7wjXW-(w/|ϫxŽ9vw[n[{m/ a¯V/'(UXMג)X.Fݒ0U]mpS-R8R&jr.S,$JqEW DSוRƸj«̦"]U \7Us^_ϮUjdSt\p=A[6L^WJ'/U]Rdp&RZ'SוR -B1tJqjѕgWJ99EW1znSڀU(t2sU9niKEV\ hZiCr#:KM6m;O^zC  jv&V[$ɸ/"$pP:rKָڠZ*Zgp`Na_'liõ#FGZ<2N:y0 ˈUO&8*ҕ>ijnZtZk/aS{g23^٧+ \@k@,~;'R\jѕr/9)X[SvMXR\Ztcm92EW3ԕP +ҕXcR=AT`,Zt0AQ;sUɛKC 6M>%MaUGWۓ?ǚ33A;"U}zËl7U8pIޛj@@:RtmXE䬩2h? q q$;u5ȏFF=҆]t]=6J'_7ZtE.o+P_t5]Yv?}]8TץZtcM(Fe u:W U+0ע+:S(%EW3\FWk])-O~0jNL5]R`1j[`Piǚ82EW3ԕ\Xz >7Oh2-uUHWAl=׮;u99OוREWsUW+LdPhjѕ&@L\t5G]ŸWpfS$ a˽ٟdH1nfgв.& ϠݝtDo=x dRɵl4@Kޘ$L뤖rHN)e}9#t`էS-O`00ac_xFF7RĞ *-zl* MJiݢzO&]Z_Wѕ2ʹz+MTc&+$Yt5]x㧳+p/uvD?/.0m>ha{?^| lA̽(AѾygsFǽE_+}ǠMEo~|qq\h_?(J_@np`.yς~)!βa6ߕs{~o|st6~b>hu6޷mpovZ탸WYojj0>ҫ>c'^~yXXNAEDO̻#N˞v85x#32{OTGT=+tבȞ?:ދA-U{O¿&l9Y}!]:ېo{gBz}[ʎ~=|E;ćl=ڛ-:ӛ_kuu>Oop9{gjI.q}4)]sl1I۝zǮ 6ař.5p )ΙmV4m;g{;>dX~\hP8t#Q(u! X|{6АpZ93NK)1iC)D&6R)j5]H\Ȣ] =79~/-Nm@n^z}j_,H5Ћ!2;K2kII bL$B!sWO!t^ИQ1{i}¹qMkk1q˭)1m ^_{'"̜?sKK(ڐ>nj>a d;DR*ƠRO cT4Z} }9xKkPkxh(qx'κF$ď|rw}nM 1d[q(iz uj ,,>6 CI-\O{s/ij)m\z]1ϙԡMR4\P<Sgm_,xLɗDɔK6nj6nEqEwҷ&MZ I k|jR !jk:%#BR c+)Mttb%YlFg\Dq3zWRLiWrBh;|CoF XڦC! d`@o%B1a>)PBd"&@g=YZkZjilEGQNڎ Gr 10¡mvpI!0 G|% A4aA3ԓGD$coba*Ng# D]SFBࣞY6NRt!}aXGSW ()7d3 Lܫ.qQr5BRB}*m糮`-)IF5H!q~#t M!DŽH :bAYxbԊ},E7?83 a ݖ,6KәWzNr>1f(Ti5Cs"@NHhI2K`b=b0ۻr˚|}+{7cwkwmY~۴T`va1$2`P8IH=(e=h^QfyֹVa[ݗU&}6}`-$ >:%@uP< *}YiU2Mt%ڇAhRU! 0阊'8$;wŗ+XqAyjN$\Pd"U̫ fMʴLw5c_%$0AɣJA&ԭ.HGfp6]`Q*YȩՏDC}^jygE3ʃ d-Š.k Hb"wUKӍ^Ovټ ȓd}E>`Ye>ZbM=A!%DB>hwal:EjC$*KtPK|̡P}ֽQ@R@"Pc*2;"<{6CܖSBEk .IҎa<P9@1^!B !ݫf br߲=V>-@Gص F!tJ35)JL T "Q8TDYUQõ`QyWBHcF8&gو^ [%VmZ4+it=liVԀʬ$޲(m@VKKުhQEx i ^IX@iT_}ŧ%e SLЍ![[&wܽW7bҵ`Yv۴|62I|U ԭGnJ''f= ]Z`w-> DiVۆnMEѽT}I(ΓD꓇FCoׄ bL=@Oz~{4l}˽ʌ' <%2`ж+9]PnDV]qnczYf1uA;(XfhTYڂDe{!kQ7Q1l]` ^yE!8ij)K:)Hvr;X._* Q 2ݡPʢ#6:XL&ݝA,+~x$mDVc-S{N낙٭E׬UP%HI21KCR 9Ok{i<5^h5э7[A[_%pg^yE D+ (|U8P(-lZi>W&Eiq#f0YJ֞ +JOQ!,iJ ]\1Fn5 q7XoCn1jEroV˥!DL0r( ;fAj$T$nBq$@Br6B!(L ΃rt `j?_m/gf/nϗ뭀}IImgw]nu+]e/l*kegqwTatEtmz$ֳbl-)FR./AƣM%M]gѺI'_Ctff 0 ttkG?K\l<_OO'?r!OmR,JbroŤx{+xjK&[|>mNU[gjm[ =GNN+96OpU` DkUv0N }FٸxN /̗; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@\'G58'YkG@@Q N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@:^'P†!9P.0'<h8x'PN#tAb'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v`tCz&j0N #y&"P*Nctib'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N vx@o}qkW^2jC/ݥXlC2.1.vݸ#B+6.K Z3b@tFU+BkgI(Е2Cz@$q0tZPZtutE{]`CW7ȡ*qtE(g:B E;$"ȯ+S ZutE(4 U[ݬVlD mݯxt*; ~غ ndӁ УYwr}jEoN18Wﱝ 4k,X U1DyG;*odjXhf5߯nK0嬞 @ַV?Jw ~)B՛EW({ޛӉdmҧ\M\|OìFדsuu/:V]M?-uϻ*@= ]=e82NJ\ F~ :+V" ]Zgݢg6tsZ 8+e4a(tC+b:B +DW, ]\BWC+B"AvRC+Bա %1ҕu?~ L1HhP.ƫ{CWm7>WC+[%+ CRWp芾vf hNW@g:B Є!-0"q0 uK턒׮B[ՁK$K+=6>ORa&{-c/-Im6_,'PZyרʋ4ƹ^̧!C:`1}16v4/6,B m㬱}3Iwec+Foባy1җf؋Nk;b}GM}-$bB?;.ЗuCC2|?-b6s~5l[Yr5-j5] A%"= KOt.s[pUNN( /Ӏ'tv^}ίڠϩsRvZ%͋yφwr‫p6lukFǮivc9r-~6~vLl.nd@Zr޽5'HӁXܸɜA"LZb2^¿J:Ŕ>m6rț1n4՛ ͱ~볤d tzh|ֶRU_N*A!UOT8ѻ6-qwHDRC}nt޿]fȐ_2{=m?e-쿯n8u Wb=SW/}.aƬ۲>1?3O/ߴݎ7tǴ<\m8R8]t8]R[y;t>pF[nQu6]w_bme+ ^)D?ա@z|NhJ ?Kӵ _:]%ؒShZ>_VN/OG))pkh%ƣ:[.SGo)'RvR,BbjoŤ1"0VxZv#?-2+Zj)꺕^SgJ JK*/Q{Tu Ei9~{LxׂMwA+)z/roMk"`E0x1m# 6AR?o/<=v{.,S_׶DOtsPAHdSu>"-p՗ڄ%>Ws+%@cUDJkU[}ct鑢lQ,c*g9Bj{f8[oQW["q%>Ws?V7:q}cw<~¢3$iRG$Uܷjg*ٯP\)5%:.HN| ;s-W#-x]ƽ(~EXE%3I}IYʉ?WR|ڝOӬuht_ru]Eޙ.{C5-:T}T}%(,: |B^IS%.eG/PJFb$Sl:߫UFe]P:i!זz]}F^T2 W S/|Y"ҝ_8 )m2JѷkD_'[$ %k%q$=DF[l.{1 +u}}hd9*S$ڕZoWAZYL$M s$K5dd*"˓x (J&3͎-MO~;9N{ڵ/ hMQ],=u:;ӏkUzk}֮ꞵ:B ^{j\g3W_R _k\jx–ۧ"TZev|zuXġmb*nEbpU왽xs*vq\cYBsQeeD66:?{Wͤ!?&0ڻŵy~:jdI$7M7\Ix%٦l9]%9 3`JFRy|3<|3>Jۮ,Wm< 䵻U~GNep>V۵p[>Cύe64T,}-0$tM֠h؀Q,ؠm9}Ig\9rJ;?fݶ;m D-\Jhg0 sDNY(TǍ`d٧8J+D KAyg9!:4ԝ2Z$&+ Vkneq8 wXKK4%4ݤl>a#U9^Uzl5,9Z8޻N$7#ܣ@@Kpis Dȅ`$a$|TQ FPhoId=ɳ*Ji rD\r t(.QcTR^iI+ T6xr)2;f&=|c҅ .JOf43YA.ȩ [?\K 0{!d/0znxSwvb4%C/c$c(H)p 6LLd-;hYך.IQpM2vL26XܙOZ>kڹ[ŶcP{mBnB iRp(o&U~[HL1'4pN6*0h hXORi޽}.oV?&+NN+ Yj޷ cnF0<14>(JPu:l}7J[_̜.s]#9'4d?3Yzl͋`O}(6#M} ʋ6giZI8f\0LX^݇"AIz0$wփ0MI7lʪrfd"*)O}LaB8[1o(vΗAŻ?*HB%`| qpI' hP =gDBh'&[ExWSE|zY?k0ح\aW&ؘC1ۜݲ|=j|=*޸Ϛ؝ mUVLmk]w#6m>hUZ߹sjp-*PKfYz{ԝJ/l'!lͶB*Pf5[v[~Teݕy0LԷ1s[鎊WA4o]{->HSs:uƋ\-j{=;;}:O8oF揤;zwFԈT'Bn2 (Yh W|µ?CM!Nt-wgO2ՓFR VYUG1+ q&"(ףD|uQ1(Lk ein]n;k]vyɭxlML *E%1;NZskh!B)AC(~+hQpD(sz#g ":ʕ ti8qEWhAt6&RH) K'1$f߼6!mNL><9-\[Tb=E\zS! 2.x2"&ZZ%ɡbZpM2-$xBc+o٢s$Qd@hd\X4I2GZyFnO^@-`Qg.1kl 2e'g\2$y"^IQfp9o*}G PЮF݊솰kY/"BWA7V~gJ!h!fk9*g-\G$Ҳ?0Xl8k}q߇,7Nx{)40 ^I?( VFQqOpt2|V&lsdG8brV<:~Cq맟|EvLO ~l)oMΧ'O~QqAKɔ~:_ޘNcOj.W 4BdѤhT&(0/H r,R?.: O~~_ bی.L@#7I[hϹ1ɆFi 9|Tp< jelHeО .xMq4F#sD&3dKG8{Ыze|qq>'u5WWlofGVgS}1ƀr$uyc1GyQ=~o$328Zzy2XHH4-0.SFeolh- @BD}-#:RXHV Hg.s,Ơ嬲GἷL2Xіuv>"JoMpxDyhA N L2:YG*A!-1R Cr,DWTTzR pC}Do <+)n@'XYP5g鬝x\ǠV'O]K*^'|ZG*Ddfe֨1Y<N8"}ΠJÄY/*2gsJ9hp)K.z0MكHYĸ 벩"bmd֜XTj3Xh*c4>Œ;̜˛{&hM&GSGIsR9#I Փ,$4 @ZgS K,bǢ$RؔFdO6nǜByT䈙Ej~Y9&㴡b jךwEm[m=ݪ1i3"ez~D3C6.,IGb*ƅ">ik^HC&fiQ%IȢ%>q$ R։ERцP5g7Jˢ]UFD#b[H%Q@SZ u/v22qNS;h X5~]BrV)0E2>e)HXYK4埭g߳᳿B{ ](4HpZH{7IU˅/ote-6s;\R-,+?Y;X0fq>ϗ{B|2N|֦1dvQ~4&糋NgggcI՝[Kt8'EИY&ch6AR[zppG~6;=۱$8$(n2ɮ# 5's¨O0~'PpЛAHҤm3gʎפf&#sm}Vs1aq eq6]sUݼ=;1nS##f׻}QJA,ah4r.O[[a]ŶgZĴi'#\UAjrT79tFU*7*2߀dNg]xb=yU畞 HZ1mq*qJH23*gh齳2kZϸOfu)ݽ&=8g|^6 ;] o9+qmX|sIv`2|&ؒW3._%KeR;b{MɪbY*鉶ۢ޵6ڭU'R$pxjdTs]PI1m0]li'QUe1"OU^$/iamnvƻNCnܳ?3n>yv:{)̝B.W_#.J9*Nwv 6.X4m|ͫ VQj_m,Ug)SjI3[8'ҎgGZ WJҵѥ်2@=2 -U8avԣ*Wi~=k)\9sNm\dAiLR1 5)*&CJN5G@"*2cʂR*s BrF$n"S2kY1ȻP_.Iq =w';{f6UBoE]/]L>6eVJbԽ7d{o([b#RM D֟\Ld%S)e&Y,Z gg',&'ǵA GL:*Jgp[oG$/1,=onS4wF{,KC%vnY6#?J%*#3T輯,W9Klr1-*&ۢbr,-J]v`=ZKf m"ggĬFFzbQ9Fsz5n:(5"O!*'IB$1 !!lڗ\NiUdiLuq$pDK?%:b'mq湯 /Fa &MԷ@IfV XREqVJeUZ`AMΞ+$7@zUQH /sܪ=H8xgG,'?Q*/1U=1+QK{")DrNh,Dȉ,4`$-$ 3MJjsVtgYj7flw*R(^Y$Yf-#9aZcTc,[A س@,y_o@^AN?1KTM!7Y7ˑAr֖V9%h9Q0VVgV4cN9zZL#1"2PcP%e(/Pb-31Nm 1Mbv~'Ll߯dL܃STwJYT)sAOl;G.w.wD(џMDe>ǴXqFTi885 \q13x!dR1t5!Ůug3:ߦf|GKeFwA6!Q[mΙʢ 5 >K<ȱxTj|AN/Ƥr]ffWJ96[l,`&c8"9b,Qʏ5AXiu8tϬ ̺RP/^<;ɖie>D 1EJLkQyg=? /aO_kiZ)8CʁqLK%R) pRy|aW^"SfH2:{|Le"`$u2GS: kHλ8N R W,NSogq8T ֩E46[-=Ys^[&uV*2F$I.|* 3]H:.{0ߛ[.7S&57qen']3&׸Cn(Sp!~J.¡ә?<OHS[b!QV40K*/u`ڤ%p3Nx2'`RsF燡)\o~w>޹I;J4/_iU%$ qRnKJQo5humyAkGZBJ)Bҽxfv/͛7/&/畳`,sT{c3^FNjIG7 i@noi%tXڌi22<#Fi0e`ŲӁ_?OIȥi'f˭4Ұ$pJc_Wvď! k7X;*UD5^$rï?;y՛O7'߿~w'|Oh&)K F~ h//\iͺjڛ6- д.{CF?vlw}jfkk@BxA|!=֧?jl2r_!gM*jTJЊ켿ҵNϥ[ӓ0ĠS I"P4c wK Q,3>]= "/DoEӣpt`)OCE>L²woW8?}ZdqNd6C9)M/jd)Y)T^^vv98nkbACʒ&ՒơmE_eO3ȳan)N﮲ԐXˆV]Sz5o~sLNmIBߙ\FS?5.C^&Gvv0{d͠e#]`֞z$WV;# \wE\Z`UA l/Bn4We;#jfW5oQڽzJfA ;UĮl1XPڽzJ`퐸"vEp"y+ ,"RWW1 9eKj/j"CU ;c {W{qŕB^sYT{ZRZ_9ǍicR<0F1,ν~}]o44E#)MNYq+%Xa9kۂjO G5n1'1@|Hx~ҍ4SwGgr+K<1D XMF`7E<iKGͣHJ[}ԇ:fOUKЙ; 8)mB ƩW|\-k7??Om(@hZʸW/lԒ *,VVK1dNn4znb?Lf62'4*qf+/Il4F,?.+/ P*!YUY3) ֺ MV0g\Ե~6f^9n9t)T޹iìb fqy2 L j;d;cRYjgAi^ߎ)VzqȄG wahy>%2]J<@\Zw깐WXqU*+⪠5bUAi^\=Cqˆ>-T\`|gU *h֋RzJJ>N: ڰs75㋫`%\Au՗XߢnD\B!YULRG];t.-~Mh=Y 0X͎N$DV6 ͇7)Lo^7B3)4)Ec9CtR[ɥLbjHLpDe-6comhaaK?/ԛa`]Vx)P.NQg w$H^duQʣɕr'=OQ_G$'n͂$+ 3iƊ$VLhi]@mwZmKֈ=R␵F#e k5`AsBLi$yd6 -1 i)KI\Rz-18-v9dFht%Y$N8TgpV -&#=貎$ߠ_ǹ,%ZuY+T8]ξOYU]_﫺]Wzx>\^N NhaYiթUZ?!!QV_0TTuK)}w4=ܟ'Ѭ+Zj]ڻFWﯼUbWy0oW}^ɬ\-i|KgMR2[_ r4\92zA#4ݟ]5RjShYE_G7$tQYسRdUYGC+@wkC 7oCO<0lskaļ+|M-Ut3Yr7cw2NA;)c")Q Oʦ,(O>P!=I s /R(L@ZdƔ7e*s BrF$n"S2Ltgŝw: \֋6mQ(AҒ݄|2-w~_2)J#&v+jNj-vwϧwKT+W rS;g{SMg&?2$#S F$(yip<*5dDT eN=t~=tBO<8r >6{8[tEI rͼK:\.{䩂<4*zO>1r-fEۗ嶏ǟˏ/諳ԖF <+өMZ wQ4jzM#Nِ-GSh xkZ1<1xe`$[L?nZfob'~|o:-َ\btqypm'џ9v ާTJ]NKG< ?<ߗ1}[ 4p>{d13"g,A#%ٕ3`N p1 ׼Չ,*GgXA;epwTUSqGBK[wNlM!j;zn=~V 8E\x A@.>&īT(3 e111iL,~qE*|pbϼ*%=,$j7%8 =xM3g6ȔpW|px ]гMWӍ+l 픨kYL"grb .ƒǪCΠַgP(L}>ɧ\êS߷fȃw)6'Wߐ\s.=Rw{gިHoe{g2>t5&t'"8<2x.E s !!nCPb'oioD43ܪE\a-G&e 3t^ шҥ귩[~8w=yxn7-ݮv[3wiﺷe9luyb}r=M+ }Y[>ru@2gZiRј>5̺v)׳ϊu]W'OSZKa:A(\3īJG=(@霹VYSq =תõkg@T> epLm|=*kb͕BVݶMv{^I*`O 5{2}H\S)CJIݓSGjz'Fmݲf,TO%4rѨU x"򒄋'en*m*AKG(Rݱ1rvs6I,Vo 냵?Qv ^W~l.N3DgBc]LF#acM29o -CxGw^L\1y8{R?1qZd6O(DdzΝB3x(SR|p#RqZћ̣D_3ekm`dMkKis=$kζu?W3TYnۄc3=@Cfvz,3ILW?_;UC|he%8EBg"4CYCN&Iߘ3Nu.{1eq͜: c)2h 4'^Ɉgg9}VFV) LFC)Gt3慰[ J_x_M CN2[ɨ4 3I;XƔ?Wk&iŽFևDKh;a.@;^|2(n.brѤgJ82Ps&82Jo@j:3G[n=mV>?  F+&6Q".0CzO>|O*VU; ,i3NKe"+Ck= "3%RQee;&Ύv}fpvj\P6g^ 1#\&9S^x*@β4jpVؔ.[J\e;!OBR^Jǹ %i80N8Sd=ժF|I1%MV@:/D DAx*tiJڐ0QEdQ8O53m{&Fw3<5CeQC"Z-hKcD"R@ kΈl)6=ɩ"l|h]8'HmֆggXӘr!`Ȓ%+S YC*sDjhtocCjjuT}/qt1+k q:iO~ӫؓHu@BliMZ/*`8}vJ(Xv\(o߸y= }J55=LJVD1y-MT|8JxŢBo ŬF%Nɤlhh1+~t4'li kPXeN:aU:+䌒E` C@#uBcP%Ǹ,֜NFXN۔D 6x$_^hg<)p(R(1EW8v@SyhCRHRh1f!&J,RzCkH}fr]6A''AE E$`2O*X!e 1y\h~k0EBX rR1EHef6x)v)A+$dgJ#u9erWrz.ѩ5Rrlqh6'N\z1#ʃ2#d_{ E墆tņA^:NyߔWO~'Ͽyu…;y_/N^}FS1"x,O@p^ܽij5MNд5z᷶+pQ;z[:ô[[cz8Kekev,:]ˤk,Aa X&3?_ iuVo?J҈*C !8~Dk^w\~%V*_d񨎒_$ Qrٻ6$ X]k{e~8!觥X&߯zEEš5AlTtU_u=:JD,O(a)B`(c@\p)D8l^SoOݎV EV1D\@p!3/P`Ơ!tvZ9T pkk+x㕖]"V>01L9'_pe' JIB+ &vv~')KcҋF2ňJ6w| tu%<M$4..4:tl KHGod;s2iy9s{"؆YΪ'onjgw~h*n0{PwI]hp=1gtߕ>NLfxvLuO_U/^.#uɟ?zbar9Cf|tB74Ć8)$e3a]*-^I*+L7U67ک#kk-C%;=bK[E,_,H[hͽs>lb,|UD6(b!-zT9o|gRKK5YE޾UTKCC"ji'm&?ȭz],CcODo ;,xGGn@C-#Lh!]mJ~KgF{x"EeUU(c1(Оif%7Q1J!&пVt޷AQ))L%fǒ{*p3"q 8m9&Ά]#BMD/];-}z54>f t z!|Ԩo?];{-nْ:IG6Y$ɭLT91z)nshU嘉49\Y#Lq6{"d0995%)UJ Y_q^1hy7xjN1W0j~I翹}O>qk~Ӳ& `kM@5IIɄB*):nQG/XJ=ʺG?/QNVy=ʏQ.ƷGޘ}r+ aci~W6q{!)TEPJD⨞fL G,AEo6jKr˹+hTjJ#=9͵OzkTͨ =Q6և;ĜvȤ&;^ }ǟfEKtG Me6&Ƥ:g=搄p6XH W]_86:Y`Zh*X#Y)ÐNόTѠKqy}ɮKG"g}VQ+sE `$(=7!Tk݅ tHS FFSR`c% Q^ d`2g=QxB]N5TH̡.%Lvdzta2]bĴsDSt4BSJ^ujs1fC"E”Ar&x**IێEp 9| <W^Ac]ұQTa@מuDFi+;2[l\,TΣF5-Qk% PeeA@v6gkM9Ѱ[ٶ.2e#2f ? j377w?<5ﺆJƣѐ*ErI91YChc |앷坒ͯLt4QeNzcTN[(Vd |ۻcQڢ݀"$NzGLTR 겍Җ1 j)XFYca x[ v)Uӟ <]`ixO1>Ubc< 1}쎟UUM9rɺG?y6? k:W٩q3!~l7s&ph„&(Rn=:Җ(m*BR.K XQ WAڪ#ABt]4oJJZ }eu5=6-b_lՁwo_oHn{lQGxpF^"2_n>z"8\=N\ek'e1WR.l0WJjGGdh\q9sU5Hi7W\PZ,eE@_E0)JB[ ѫ&(b U˽iOo~\1}O謚D,ޔMo/"äw>mxY^IXyLG`14oJջ4WlIΪ%qug ''+q2Uw%*-]dX Erj쌐4.5:jVKKK<6jcߚOy2 yU}?,z=H[,bJ."kPg53 xH>ds(ho5Ѹ~vb:?YЂV%]kKզvrkgy^TO>a"GՊJ V+:V#)-=V{XM!7LrAgMߴMuFeYқ'%G$.u1?ixFALv:'jTc!)%z9)[8`%D"4HsɅ^,5Xm2j!9 !O)˙AK9br#![HP.Yi|)wݼ3X ? ߇ë$L,#\o.qBfljսy֍]y&iz\=pW\ҥwsIz  t1TZR1j8z  ׌-y;OF;vr4N~eVSV]} s!^Ü\WLrY_F}Vu &_'|#~ex_Ok7M}vS__#_7Inh_nk7M}vS_nk7ur>w(rgQgTT;Z_CQUC;ժT-vm>y9\Vy= X;]Kcc8p.mcGkR9sr1[FXtāp٧R 6jKr˹+hdjRc'50{Mԣ4Na"f5Iƨ =Q6և+ɜvȤ)>?>bй(Plt4ugnuZ7G Me6&Ƥ:g=搄p6XH WUG{`F'9 L ^r91%c@陑 !t)oޚ8jxy,FBwf E bKVk:^pmr{SyӟP*V,5AE% Z `$6Aң&0ĜCRL@[?ܥXay[ #$*W/eGRQ@ERSl$eF6{t$&AHShr<%͔g$1%bUR09!:fT1)8Xn YD^\e(ƧMdf"PoѧhPC9$CuBQQF+r,%ҳ6FM앷vsyxM&nE}`ݪYa+7޾y,F?l&FfO&79{*Vz"oRP'ysm{ێ'x+I o!WR*Ad lV Nbmu09 'ml˰}:K0 r߻Pa-p#O!wCi GcBM܃ ?omTD(4qx71tͽ8#0Úc4uE4} KyCԬ8XQ5v̟#Ǽ&=q^8Ygh|>Z~KFcM.vx>,9x8t1܄[ZFW쾍Kio>hZVk~ KAz'5յ^T{)QZD,)3?cRȣ Ss/4&I 2F#AZBaBNm7(!j6cQJ(KR<=SgG.7!O?4 qkHWɇMxxjkvW.>pZgaz7}f*-[VݻOM8jk;E{hDhs׹f!e [غܺn[os4Eͯq'~h6b[lݽM{;wy07894wQ-rꖎ<%W]oB_45eyWnF΋[y9C&<>[b~zvewV遁TP<8M(QqiTFH.b$P/O1᲌ǼS聧 󷱇#g<!!M*X`MT@C¯$4oGRq ?{o\O^!5NL[",C&BIyhpLO|L9ed&B굾dσQFYWݕ  mVew p]B|g]Ot~g9s!A &C x+R5pF+,&Io"6+ !wRy@x iG &Q)s*i/@,oNG՛X|Y}D=Q҇ijEΘQh~2TߜMqF3T_Oǚ?9o?|3h~Be|>k28=Ϡk4|7MS7( y5_ʾd=1NF slݸP.X^;Ums.J]r[i-LayUF!\G}-lJ !rWX_؛`kB= %m5'7i:ihgcwʎնgy &Z ䷵"8݆㾳xJQWs^%!`?٪`=1Z`V7OqɫfZDƳ8;~<ޘ.͒e˶MΆ7=x|ZX,mv6]4|bhTl46ochw\%_y5 I㤾77a5PB {u|M#$nD4Sj_;1Qhv6xƭR&٦ȾDcB=) \7ql$5N^V=u+vs_8GUkWn{/a8vѻŃms6aAE2 ?sK!9j}\ x3l<=~/eoG ~uڝfۯ/Ӣѧu*/yg]ΖwsˊUu#Wx<f֊fkj^*\:['[Ez=gl.6wsa~ָN@Ma 8Dӌ2HMYƥo|K :9oZ߻Qvˍ&?)[Lb+䇴SXj6c{S`gF'{/o-j fN`2=s^.Y.xtfv 1!d3l9`Y;Mܼ3u|/sa/6/<ҫBH SK Pf qhr\3JHDQ0!>>wsxk!k_?-w t)$B`RQTZAF+hD; N^!3M~k͍Z~rIvΙƒǪCN!7b(?>)>G@?>ͧ^4&ZODE9zoԂRZ;a]mqPFxsp9>eӐ(kŤ>E]0!%x\[J g)R169f#5&#`5`UqQ8!('ȶ SbO!n|p~:YPYX>٦9wڦ۴|Di5I*i/>_BsuxZGhd^ m9Ϊ"wvK10!,uI'ŷn-՗Kl-ͳyzpp*,/=z$%4UdoL.XL Ш5PjNL2LϺ>+u}b=yU斞rV!:2D!`(po:"h%TIn<ŞkZgg@\&=@ʼ8k|}׈+uöLvP=E`{bK-ܰn(JArkjҮV4}e8=Wœ+;ɲ9ň%0(jJ d"%RXSH4i4t:BNG_d5}@o;>]H=7qOV{H%™i!\?qR\2z@ee: 2b 5 2%͟"8hQI4C9TyĄŠZL f5WV=Wb䩻*/[H> EQy{Z:jIheV"Hop#T-1P)9ER "2:Ř'y[ B,1YϢhRN2.2GֱPqbXXlf Ua,T==E ,35. >|8[?m {8>4b e(Imy4=Q,OܣF͡=ѡ%MP;nxlqa+JDl m;blb‚*H$hSxL̃AR㮨 Qgm6•#ń`m" $XDmcp9P&$H( .`T̐ XrfP1 8$Q ڹxXLxP cӗwEDSM="n?ĽΩq89s\BQJ#O$pR4`5И(k[C QZm RpjҔG-X8&GQTs0!%D~OBEn#J#.Ii,+.¸h{\qqӌL.dOBb( ޣH-e*䙂HA=DZ}aqG<,va[2c>5rnx ~AbsqF8A{?>2K&*@j&g?COc~Ф\gϗ9H87y~9K8'x'8O?C`MU{ep6r|_mSOE,4Ka0Gh$C%W<F:"΅?ˏg߳A_ 41!x ӐЃk ݤ]Umۅo;"]>J`"} W/S3]\`g;hւ/P}ꇪM{6m U a/hLlz:qyZuT>0坵]xܥ&4Ӌ+ǣC!N# xmI&I)KݱCJ(:m4X{`AiDzAR FѤ`ɾҨ6&LG@xm(9D1%b ogGwƥ}hhl.ol9l[7B+X7uڎwI_ކ?F{,ɢ:Q6V.5<$6,K)Zf: gf85c2(LOJ-&v g}z<i'__vm_7z{0,-K\}c^cfG)PX-%Վ윓 cq͸Oö<8`Y8زόmNMG9RDB%1rm"%ٻ6$W:b+ Ve݌ЃLk֎f"FJ˨S h8 4I@¢F+*/2-)AW g.&nV/D9Vs}VH.:2B )kK[׷ [#?ԡ;ܳNѾtR#fB O}fp5q~Uwc&ocs^.P65(*kcSrT:+1nj Lh 0"$=- ? >1 Z$.cPmD>N#%T,Z3cZ3Ke-4"#Gy$tT* -pɚ.]R2KH8$ߥP3/Wlˍ'+4 Y|EʐdRJ$w3Q>q-hrj!$kJ߻ u@=3 `sAqҔ[E!J&(.aqNJr؏CUw; amT̉vNAhi^Ӂ*BC"$F[?3`#? 45*e>iA[ 3D#NEcYCR]A$IZ5 !ΔP)' 2rVJė:`EKy??b}ߵ @GJN(&hᮺ0 sm9"1: È1Mc`O, K㋓'n\fT*~Ezył3_\wFg yӹ$J2F(^޷hJ{~6$Z=Dgcn#'Jjo ]VH<\L>|u~Ca~RE?oťtpe_~>D''27%Niot9{5U Wh+sM58ō9M.&0#Z5WW:8M!Ƶkzݏ瓵ezmk"dq ۋ#?V.V ++3[JG U̖|}-zEw|PCݣ&jԕsͼ4uOƙC}5B?Z &noPcxDx7yX?sd_~}}ugϧgo}V\q  IVp7͇VqLChss\!BS^1oǭOvܶV؇WKwsoנ۟jW͠+~(+H6ŤUwT#*HC<_JhUqm}^wH7Ã:J>sG kbIP,oh2kxbtvp2+@fX{e"QA8n=Z$ӈӇ;٩x;+-L:g.-UaPzlS5X/Bʮ_ qZHjbʊy*պ䜈2].kpĺ$mkr_=heI3[ʍ*F4JD| (6|XDipT!JN{ϷEzvN!k}^._3+ K'E֗C}EjJ0t2ZkБ=꫑ Vboj }髑ljd*j<þR1+$Xuet_Uve*>gУSpmO]:~NVd 0F_E_yY9G3b*vK~\)Ye/cDM??NopMQZa˝B]Iݜ|S@')ޭDi!&2W/'oQmn~avYSܿA4i7f &nޱHkai)O%ĔN8S&(8unO y\c$?v . N:n>=6 57G-HQ-T-㸭]v'{Wm D^j& ^;8*.qfa{5_ZyZ rrJߴ@S50^Jkx`0z)^O4 M//[cjhAR MeN~_'-z5 r[O zN0~F5źI_'d[6>E;4ZP~䡣-##D>z\F |ӊ Se^ΡP"4Jk VD܊r pW4ͧ>r<쥚*l2杪0t>h׏ &ds2I ;\BqŇϊM4w\ \ޔb?:^c pƾ[eёܯU=9kF oۃNsKYXjF!\.Ԝ2Ц1rm\f ]{:7λ(SdQްEeNre)cvB*cS*Wܢ"h_ʯ)&mhF#nb+|`>'X-ssƼ=eSsEŐ]SoΤ}`0.N91pƋluFyZ1@Pf\k٘,׸z.ΛVViuڜ8[& p~Ζ$"nRwpЂvRW[ wV'0L>W!AB Y}UgHq90i򈖀 LJ H~l ,%]+kb3d |"9{HHG$"SLDT.ơJD 7͗tgrdH.CIJ2J()\`v= KS`32 O0I+-ASW`F]!}QWWWJv+0R$ݩfB~XHr9=uR֎TG~zf 8(;!5b8On =ܜwDIiQ'pAo( (ZCWLU^l?y8> $ 8InZYnЏOd^׆'LF#M13BpbG Egy̩ҥP$KfT>u//BTǵ'\IYlpfYadoFLf;(ds6vv;F7>HOG2rO\-œCjyO.SIn<9pE'7[Oxս&[G%ݱJզ[ODK7*a_R+%*SYiqPWG]AUYjť2[G}-Kr_^ 2z^_ƹ TLIRRN:}1tp4Ѥv#͓8xTJ"z}v͒Ӻon8~T5=gJPITR*3gfyUd _0̋G17s|{K!a=e #X:G]drx1Zmv `s f>xr2$(9WPj|.M]Z]d{e 6ͺ7\TחW o"l?-?垭.Q`i.pE-uQFGs"pYĬF#i*@=I&XD(IolV!=%)Ĉ6YPR@ưKy᭑`v7{捷K2(IhJ\Qbgm5rcFMhiOHӄh9HIJ)uDNsR{t_C-u@kN 'oQ.ӂHJ;Ou>}s4 /ѠU6\Έz=&դCrWdld"%Nٔ q>b!PJ'Q c$c{QΣ2UDb̘|$0g "4i7XD)I L]2$9,tnj'pIv",AˡY.aSs9r|7V}#\GBk=WrOZo<@'> }nlҫ_@PLH]U*fװͳ>3NjܽwA&pg7zj"3cn1[A(G()pJ J\T YI `Ѻ~-{J6omV&;\=jSւ"8Q AdɺQ2-xM},~D庾6)o-Y}+lBTtq|\fo<; C>%Wc'8]zͻh|3Z}+,,eSL躞mujfqwH⍻ĮO,Z>#-&Kq;+KZYwo^*2;_Y(V FH q,CuăVhZ1y%ZlBd&> qRM:| yB)@P |2 8D@i'2h ITLD!xC={i>p`?"tЧ{De^ȟuϪά 'We7Tjݐ䣧D1|TBG6 (VZX[PDdg J0*<)m@DPnOr}j;=H[B@qʋdx@&ʨ(f pm0<D(n`Z$bx.qЖ eV1'XtGd%b b4.TN?'/p>rwj5/% U~(o$$#iSL^(W^`L[G$ަuNpv^{O:o X62̫1MXx 8ѡM[-NmPnmg]+J!fy1¾RYZu+%RP)D˭]mWC<[)F5(Xc2ʓ4JX*'3ydwɱf%w|˳fAy( DXw,/K 8<3[nK9<^^BN&en% NEXdg2PF,7J5%A xzI^:l4dv-2u=.qDsq4UsV"{'HPUVW40`i)Mc>Q=,cH!Wm6ek, 4$vFc$D6#Nۢˈ:"YJ7^P!8mEY"(NHrA-+Ҟˈh'_ |NY(9N'w͘l:!qåWrfgxUeaw i;u; d^{Ks{ U8|uj2~|׏f!e]f [Wغ~ܺjZo=vh6mޛ e-V[wƻoޮq畖a:o⻷ϼΏGiͥ;n.w:Ǒq穷^Mg.1zܠ|KhX!q4n)PErF|9vW |<>Ϸ|;nzD]~Kzb  e S+LDr@^/ ᴌy!"O23/NCV&aQ,h HC¯$8/uTpcGRߨìw{qG=cQHMQ&ʭR@% ;d"L2Dcx!] FI|2Ԣc?JLv RTڲ\/d!TnݿhξrMCMA LC x+jEwxd0I0pZIv/wǛaf׿ap=>lO==C청^vp&ԏ?7h HL_DA]oy֜wͯ-:;"ZX+ /G̹p)$B`RqpED*R*)D V"Dv<$jA=!W8<:M:Ëƚ3 _Mt5q\ܢ,/-Xͽ&Rq}%Uu<= 'VoJ;4yiy!;xPm8R7h"<Kl$}ԂW B2 R;y\vp-Sܿ_ee*߬כ8̱] opFfY`~5vrx݉7ͨeٿl)QYhſRiaroKZ!k{ W&ڵ'۞i=Ӓ\{4K(tTJb+/@5=X}z#"ZIҚ V*'0ȽtJTה\ ׺q_.='=A])w}j~kbSwmR}ˆ{, Zd19ʡJ$)* #mvfBNOn4v:{ $*RKaDQjJM$!8DJΥGVhi4t.ߓ(@d.O6 pe|NwF*jqK IE@k)lj̴lbD[FTH6D ϒO 4Dqhr#Ą8j<֊4Nnq 䥇~פܾ!EU,ʻՠMLB+ôAzC%.PP:1pSsU/3,EFetHޓV@!` \p:"RQ*s2 8W)4P,TPXxeb:3;>|8Ar6;)H&g 4 G4&xf"g*a$ #@Ђt=.OiǁxIY+4.%>ո|JIί& ζ_53{ᇯkIs@>ܦ6]'mG"ݦWcâ ~nWno3Ԏ,^5hր?4h^O]~ABԓJqy|=u`:qy:hrDq8 F[kOC4JE+#󾲄QK01MO$ʮcʾ3vNuxZ2B%1rmq@#X}L4MhfuPFk!h TB%L]SThf3RE֗7Tⲡ~euh-A1gϵwcDC_-.|PKIg्TQ<(Yc`,`u@!"FY@:h>=.7/LJj$lG)\\(1kgi٣  / uFV:S1S+LIK=,-,HtHh^8d:2)T-GNG#㠊4F;Fd7`q x#OMgД p/&E zFjmJ́>'2Zi# SMqJ7ZI|a[?o!`&+ͳT2%nYF(^!A%Uͱ;daQȮp0qzRkpw]蕪~eYW=ǥkCjC-ظSԧiʎ'ZK6Hx_q"@\o[vDӴ7EN8$9 K&W.=IٯW\O꽧Lkx?݃5>лLI5%O83P6XON;97C8qI􃛓ɔIjl7j$ҚƒwQ&0i \4ohBc}QzxPQN |v p bXXV*0\ yiA%kTY푬c4rsP1wN D.=`Wu1L/V x bQ B_u*Χ&0N%)seb%%SU\y-D pe} lk2+Q;_^9uYXr> YٟٙS܋KYȔ6V u)ufA:pEkD>e61xm 1ȐQG<7d8M tH9Y y}A\kg.;0y~j^5Y ?.sc^CȮtr쬣ylq҅ ,y⭑thHCS5`b`@ K~}gFUY͕̣'J<<*HcBce0fRTvelJ orRxZK *˔\4֙ "912:mrl Ά;cBu5}9$q{Os]qla>4C_z!Vzz1{gO?_;(̳%v9KW Ir+yCL*ˊinsh*NsmDdUple sr,!GLkjrcpǣ4;"]>k-wʿ=S+:eu⪠41+nyXgֈ+8tqUPd(ŕbf!]/Q`=ϔ=SDI%^/|V˟:^{|8L66xVc^B)-)AV k\Ȟi\D˓ -T}TxBA*[O|;0dwNˠu*$* ,V(YDTB:O/ŕ9/ъr()Z)x-ͬ_Igue]̤Nu9!ke4;*$*ǫed`[MJ 7 F `OUASj}\Ait}7p ôxqq`!/Zx4Lۡ&tjץ`d-WG!l@+cUA;quJ"qEQ`kU+U[UAR1ۡWG(s#ӲBIAkf~z*5pWO39}tsoN?J{P N C :7?@ׯU;y;";dI3lZ) `e[) \ڢ)./{蚢Ǩ)ѫ*cHÉeLQ cBJ,*鼯,r;e1zoHe8:#è<`]PdbDR6@rVF̱2N8J6Ld)de@"AdbxS2dTYh!|jBPsFF[Á(\А12 VXamIlt,$ |>$7{\Po _dMI# 6f,-{tAZ\bXNfg՘O%>}Es ,z 2 DJ"yP&1 wJXsI^_)64/Cxh"}) L:xDfd;Vvn*aq x#OMgД p/&E zFjmJ́>'2Zi# SMq>WГaHdW 33&^Fl< ?K9&ѹ2` J4BP ɨ *y:h=XFW7QȮܡuדX蕪~.Unco}Q|̮_ʮWsW˻ F9,5iHL0tYxFoxYo@s0(8$1}OVKKՏi8OoN]|)mF:VM jb:WqaNϬpJw I]O_s6) M>q?߬f8}Q̞~ԣe/׋oҴ 4ziy(|eYaZxȑ_ضb 8ef/ݙ]Lf&dH![G/S,ktӹrB2%7ߦ+ק˫KU9nrr:NNY4\m~&]qA*$ n$V,=ӴtgR\λ7酷ng, ^b2ގ킡Z벐^|p{[i%[ llllfq'A?(2E@n^v[LrmVgՍZҼQ 4Ұ΋y=+wENyuXtrfJ͉O*';{uA?wo}]|?7߽?`۳!q N6t> ?=GvNݛ]747kZܡi M>t+NJqM>vԄ溵 !>t~~; |3wʦӭ$V-,؉PI/yôWZ H],'ph,qdzI#O*`C6*Z& ƺs?NZ]h>GZ+/=_'ʥ&4~J?-.vGe%BJjƣ%{MTdVi~uC!fol3Ta|lНiWнʱEGτ,pK.H;T䉷 iztrЧG*GF2W^av*όWiGju^\M(j(cPpQ<]3, JN%i &=r4ɼo9Q)&dƔ7>-2!$gD&2q.30rl;;Zޅj8昀Is^ ޏf t-c+w=C)ԷC7 Ζ9H"m)jEB؟*bS TeY2mJp\f"MΪd9W:hp&Yq6{"d0995 J8Xӡɍu~ ^9it~ʾSr h0KdK6,{t{nmSoܚ46$+M+w@m8ea[:LTrG^/J"_&{Y=v\A9́<@$aj\,٣ы/- !EGPU`1_yb>8֫)P`1Y*m/\$kq 7+`<YBh,e֬)@1kuR$< d*cV2":VK ރV pjuG7>F43۱R.[G4/nK~"+<A!Yz)kV8&K0e*LNT^gdelAPCys(F[*IB0pV~,| h$8EStPN7FAט*5#:A6jMHs)!wiXwA&u|ȣs'?dCZjw |Xa8ƶݭnC|].>GC*Rv:We+r&= mMlᝢ/$ijcmC%߶FMJ/`Me!9HO) tvXROWWRͫ;7zi(tt9<Æ 8oLC]AB]ƺ\Ad4'QQ՚,1%&WK&<+bQEKx8i]T :2$S*CLLYY_v 9K@+tB~Ëі飪X1$e-hZO@]Bg|_-!g;l؁hVlbϛpࣲ2W倫 Ak^dMZyG"}T_2nt #'!*@H(( A( Z7cf[?lUWuŤv]BVyGik5){ɓ^ym\x;v^{#w5klǗ8¼5?ڵ9M~TtDu7rgV(}_/ϡ _ IYë0j^y42r"rfook;sP XR! =2y̒3d2FXcbҙHQPSr))ȆoM)z$Lޖpu/#Qe..Vz=AmcّNsgh6dw+:S5fom*^Cs'n_\V&,yfzwzp5Q8{;wrkF1酪+WK脺AB.os>tt5)zJes^oͺB*YQj:[wvީOwyP\Uϼ˜'κ[B[Wx}%$j?ޒk>5g?+k2 }zmCrxy޲Cv{l"skpsɅ{}}6 8Fd.W9PHVyh)NRde_!SWXf/ _j/j gz1 ӨAb'eSFd*J T{τAB7g^+c2QǙ(fAdi$%ݬvZ x//  0)NKks<zw%S a/+B˫- ۅY ДSZ018U$IdBY4!Z$ FXP3Ϝ]֯ͦJ?ɇg{6W!&PwW 0㜍KrF.~ʓf]_59F(F-kdSN 9bwկxN85؁WܾOQޮ\ >1P{ e4(VNe("€:N";'miɓVLD!*lrkw˄FxwoP:d#]GP">~ 1xrV)?I?3~tЦ.7ͽ09AX0߶QP'"sa݄(a37t_޵Mu'&avkWvx5@U^,JזydEW]fyVW8FԬӖnqkyJT48$.5 =m*HV-\nT\o *ϵa^u.'YY7Yp]IFxtJ.O1l W>0I.g6 +fxM<'h-r=gf u3݈Y DvJ|= $AbCraZ(MƠ)܀J =%;t+>߱"zV 8 vMF[\ؔ 4:Q*cʡ W__:+jMMˆSqL=IZUPQ̣VMd{O3Og!@lW`׆sp{?vnp*Fxx2ݏ`NOS,ޏ\?}ݽE%vTirPG(%=f#gDιߍwZv{߬l#o:#](* ]psT?O?ok`umBug3l2o^j3}isG%*/Wȍt9cM@\U+#?~@׉ -sqFϼ(Fr_ NEƵsJqq"M=FrT$L5(h> B r28 ‚Z5C/:g"ՊST+ކ\7tzRP !&0zT T9(\P)ŢȨ@s$S!qzRd 6HIXRR*sb󥑱9#c9[mPB5`gӞA.2 unOnk~6 j;O?GKrzʣ(v'PBƁ@͢%l#cvl`;8˰ɕJ8%f&脾0Q|Lk(ȹt\̝ARܱ-j¨,}6•%hL"Q O,VXE(od, }@-C*ЋQetJQN<΍*uVC¥Έ7AXg%usI%<4@Ɛ}]@8}BŃMn)[jǭhM οCJ14E @ي/"EO'Q$D.N5kܠC,1 FK )VrK`RkjcƉ˯ z,8eT!2i4*24R as4ݥ-VžȞ?YKwKT:w(.4BhCI)ܙVQBLh;oK,cGf =[dU1xϺ_9:-nx8DQy!F}Ks~d$,}|rj&_|ScnBfI{Ȏ^݋QG/}y :XDG`7 ?ߋA叛ݖN*14 Al0a%z57 ~QV{}su5iڑ?׭q~Xٻ6r$Wtw؎"/XػOA@HYJJM~_{Q[n漲C aDZ~f6Oi1EKϫT)T jEc 4 z #Ly'vtMy [L@$'2AiJBH!'B(thRauqߪ /jteBwnO;EdӍx@lFB51بkK>x:"C8>0wI>t`U]k,*` kiڟ@Lmz k֮P/umF(Δu>Y]:em)U BLWWNèHYGȹR,t X YlMy^ (F P[hxgMm4QȒiO(sl3K2szyLsܾ>)bF̒!8cedu*eÿNU85l7u̎96q =;^1;-_dB.n>5ҁNـ ^Js lj)&aJBirrǷ <ɕzwO}cl`Hbr lSeTZ=OMImѻ,cLBGr^xCb"#)J:1'QZ!1錜cp96$ >pq[ԻmX5']]'mE]7!أBGKz U|5,*y콄*(^Bﰗ5sc"3^f76u ,F,y'9f{u pYAFb쭹=u5GW]W"sD]Qt-DP6g$%cCQ6 +<N#asy-hr%]~\`5.R-?Go"rp*KG$B(wlvFΎmRêd>'o "?vKUnuv6y6RGbWF)fP[<3bvU&ZzSϡZQFmtIKt1x \%fKJZLFfDו=;#a^OYt ~}n)Zww#2F>~o_.ݿ7oVw*,4Akw$ 0~,I6" M-)yպ5hm:: Z`_ֳ`~{ۓYi&Q[ڡ7aaω!jwWg$t;ڣ̻9h}>>efQiTYbb`"Q#WWNy `Z%CѳMF]o#A7v0igIӂ;mo¼ A ԅf1Ɇ";C 0@S0I-4TCER"ANA`̈ Ytq(Mƈ%e4RQ'JFt SR$ahdx2rvNLQ8maȴZiy}-+<]Xs;OX Y#d)2TOVem\($/DJI(3d#@,%5:LۿAzN9"Nq{jj ɢ9%kd .( F}ϯVs(L" \ ŋZ+$-!1)HJ'ձw>%`H2hC) ѬvbeU rJ!s z;&j=y ں.DԻ_F|~9iWa#<H10&4 8DlbЈg;lll,kod}(X.*ּqр3Bn] Q[!)dx3y9ێ*vْ>=>{uGJp^{BG/WK9Y{6E_Ze'kt Bc˟GAP_eK&)8' &ZjdSLI;`xt 9mciuQ HϚUVBQ:SRP!u :#a2q2>.O.2^>E3o+\ ܶ|qh 3/?pIp̷ZHs%]Qi2!y&d@SR`yngNwHӴד\08Us \o{Nz /y\7I\ mI7尋Җ5ޢns0{{;9v+OѩJ+ 7ԝȹA^CrjhnBG*e)b-]W:vmYb `?/%oWxƺ~D_.xX:Ur(3h PKQ6Š2%N7Q"2џ̷c&'c%v)e@ǑNouzfNnmח]YpW' Fe_MJ.kVp}*l' c{O5}Kȸ;m&jW$!z)}رJGTK9*ɃR֖J*(H @$BjZg[H|69l.:sBK`k-R1L6xgH[(~= m\3:o(.?*t6w0J_?t/J|9EPL»O=sm\ŽJt䶼Psݶ}:7dTI9GG7<}:>=/hc^qjȆG6ݺۖ7|a0Cmy핫ny=h{X7{&^q_Iڏ}婷ޚ_ݵ_,5wxxNl7m n[n+\Z4 iѕ3ES=iJS2F+M4Eq)#;1\?:m8f'ګZ=#DJe@Afs"crLx5Š5E*ihuR~GZ6ʥ+rj&N,5Bgiפ-?2/B;52a< "\2՝i׿ ^ONO^Z^|_u_Z`v"Y qŸ~x:fzulħ f-7VE4 |a#9+b >՜n"IDiɸ8&%)΀4Q}-F搛6˻\ousݎ0Yٺ_GUYE2T<}?ػ.WpGMsnMA8? Z`2<򪚊ɸ`KumAKDQc!?o]Bcc$L:_)RySbhjWj~wO4\xHX"m˞t.'`Eg9ˤ SYaד='hm}0|1Sh9ţN[JD޷T;pt` i(,AP''e񞑦.6+LmN"XBYWbۋ)`w xx>:9WtJ^]xM 4&JʞkFEu;|YE X#_>Mdے&.a?+[V_Z${HV:&YtӊM+I=(-0_#o0uw2|4۸5<O;޼]33r£c=>|ҼR k'(^+a6yZ_XڢJ6D/G ]-:~ڠsm{kmvs]JYi=B+V+n3XyuN{Wm] :QWʞ¿׷/NW[ \7q >ph+_zEfݏh7|2Ԧkjj *,Kaь3)Ab. kBs&MAcƊٺ\"͋Г}㖮6B§[ӎ-BAi`¶Yc1lq_>Fb8&}$ 9neyfuXdv%COI$nJ&IJlgm -j[!9ZqQ S}'_<+Jj(9Z W)^RUj-e#!pM7q?97ReSiCG2<2/o)Ylm.WK_E%_唼xqSg8|B㮝Rj4YJ N8EHJbVGv K ǰe4;빌Wd0:IqTVs\Xj#bSvI*f]|B7#|_O3^՛:*3zVJKIo/RSdq\ZHZ5\ZB䱙q483xZe4.2B?2 ? UW]\_|`e?.poXijN/tٹVFʙN\$p`MbjHBoYgIƕɸKؾDd/lJƚNs<6pCn$4CR-צ2Ȍ=;]hX{,8yd晵g{gΦʛ\Ch: Kx{\r҆)s:[Wǝks!CjBmepM*{S<#8 /T+*#|8polub<~,8edFgF`XW!R&mM\[lS @ôK% X4MAR quduh-LuI%e韣xͶJȤ I4mdF C(is4+9ȼf^y.OIOP*5BR( )!MAעS-!+dqO‡Yǁ|hո|{{'{NXZNc/]ՏOQWKhWcvYX(RXdꂴ֋H!.B[D?zp*RYx;*F+ MqiໜkD(堵XcJ(EQ|Nn'5Đ ErJ| xcyH/)iI^yl;-w/߽RM2w??իܭ^o V]3'ҫ7zl]1"e=-Ov뗏C^]m+"RA9v-87 7JzNXz*jS=^2oj\Ee\Xz;wMY㺓ov:}|^8Ry2blI?1]~hE? ]2qѕAW2աC'2!M:\gBWcygHW̄Ϧ ]u^M:Z1NWe9ҕe뺟+MT誣O;JgztEhBt ]ua2 hcRۙ!]9>ӡ+$ׅUG骣܉X3]=򢬟T{u,\Rjh;v(w왮]4w[ ]ȗ誣u骣ngHW"'DWUT hYUGuWϒeWϜ(7(`Fz7XdYgQ s@I4,yQp#eZ\-ߞ=EAo.bȺ7#FQ~H GvOC{3ȡ9%j5 ,oELt5G|`=L.+5WOf⩣5Guv^drC^*`ԓ~S.?І'AʠmT؃LWM:`&CW.M:NW@)gzte+6'CW ]u⏝:c;NaBW6fBt?Ͻ^ڧZJfztE!0 UKj*tѺWW3]=Cʨ#A`3wUGKG?wQlѕ#a7%r]+uttztC?[ݬ[VԗzqYo1!F(- r%rO?|w-Vgg9^; ^E7k[Ƴr{ #߬vs5y[9ꏁh D?yS_@2F6mɦZ~6?hb8=/.EoZY/MRΚb=%YBl/%q0)&R*!F꘭U(6R1X,n,c>0vZ:[;cH;0L\ qm TCa0&64'6GBAkO!?&Z5 5$M%c0&ڠ%pӶ[d Zt,k͛sH![j͹\SZ+bMPR2ҽ)VZIR`KA(1aИe76eZIPt)SVUB)F幰==a덬 4YAېRb#[[N%caM:F0T`H^$劁W9Kk-Ƒأe7Jj!ta b~{z]F\P$7q[z{sd%,1k넶ܲ  V!E%% W,|œc><7!UY7%X[?<*BI`UAT$ׂN(# }gt10z7_ְcD8:AЏkc6b@ZVqm@j 7&XS!0@*`!!S`IֺK550D@ Xڎ QtFQM8cV-3qb+R>A 0jjA1 F ޸TDtC`JڨȈ Wh8CZ;V@`1'`bB|T9oxAy[8CAY J"ZTlKJc( 8Yh1FPTPY.NZ#'_p{0gn8@G>+J I/Mҳ wSG < 1/,RԒ"4A"e$iyM!}`K . %}"y&HݗLd[{#fH܆m ⥻"Q,TG7"?bjd",9 F][oG+o~!.Y1&K6`ԐbgQ(Ymof۩st3[~>L3>wd*,hb⮆X-B'0_.-A}L!&*f7]/룫hFYuDh0ePwEt6yЎ2ƚv|tѫ3K:vUM9ubhdEFyh|F"mQ:2G.b:% gh py2Z;xwGd !AesWcDwE*3ۉTICXs .wm5,MfgB޼jXPYUۭ x/z+^^a;-Iρ$>WN6K `ҹdݦ6`-A+[tiqޕ`itmlZk3 =?.K0 馫26TEѹi0BKQcEԺZC^kpyjh Ƹ>$[C|?5kk8n0𐗘7C|̘ty%\ Ec˄  ,00 D wBA)W;53 j+Ip |rp5rʶb?" \2:&3t(8q˶ %*hI%@b~Ui0is>rIbUr "K1av̂f%d{gAp$"~ŒPq9@, v^QDS F8ØcA(\oj_Oz/n͖}IquXIp\2_|9}O;Bi)Y]s%i z矿F8t%)VF. zv I{gѪH'tL'6@_onٜc=ž0c|pvE;r=UOIkǻ[7'G(rt&(c_ram dMsЙt~B1*hfb8I .GJI|I [.( DI JQ@$%( DI JQ@$%( DI JQ@$%( DI JQ@$%( DI JQ@$%( DI JQ@zI 5$̵ & %sKHzI $%( DI JQ@$%( DI JQ@$%( DI JQ@$%( DI JQ@$%( DI JQ@$%( DI JQ/cPz0I Jؓ@ 0zI I/@$%( DI JQ@$%( DI JQ@$%( DI JQ@$%( DI JQ@$%( DI JQ@$%( DImPJ$½Lh>$P4( @CP@$%( DI JQ@$%( DI JQ@$%( DI JQ@$%( DI JQ@$%( DI JQ@$%( r@nwk^4m[M[ewszC_ܩXt܋C .U .\Zo>g \"]0gn0tr7jhsUC8K+o; NHm~]kgݕvO}vAY9G h6š߾6.'r<-@λV=^ƣ^\gzr60MWp7 k-7jmPC/:N@70߾ .,\|]Pʣ.b`:=X C ǮJ'I75tcNoǝ0#~<Vz(\/~0Z2F)UCt91Mύt 鏆z~(t:qtPzAtJ#UCW=UCkJGtJ/]5V\'BW vO+TGukѕBb@t[?jO2`K/{:]5Z]@V+5$j ]~P h1tj(wf]2j3$3h`QW 8; tDW/l 1WUCkUCii%ҕs\ > jQW /f^$]y% c.Nu Uߞ.yPJ}t&oby 鬤͚J9ۣYU_LꫧK>U.7ً|iaS4XVcWE,wj%DŷiVM9Iiߚߞ{s{l4.۞f}e2v*xkljNB)LSe1"kQ]v*1E=x=?Hh{19!`c`\kųGW(wЍ̦'Ճk鞙W=3]= y&zJ{d]=sUl`Oz]5骡DW/RU`}-BUC +C2 +𗇡骡ɨ]R;$uU+P誡U骡<Wjq?+oh8]5j0kW >z3P2/,ㆩUָcґ|tb`;`624׮ۻ|tvc 놕`'^+Ƅno7 `7f&}0=~PG4i q=I]?)R':uHŎYǮdpVA{Vk|o]wǕKt9/Q ׯ>T/Gv?Yv kCVEr:J}3P %(VJEx*R SF%JppI̅Pߖp_lWHoEӰӞUnMbOqm͔l4}M lrOnk ˳I~ݗ%QMGk{qNV]_֟cY,s 0?:]8i{˞JHiS{'\w8 C;wA!q.FLඖPPٷEE:qڣ2%C%k%ѧ!L2|=hŒg1܋n VG`Ћ;~>w5/O,*s_Y Z~z +? a^x)L&^lN6gmuY~8@Ks_+v)ƣ7lXOO' [%Oq5'5\~٢h9Y%f~ 6_ďPoP+%/t9mm7~VuBX??h&-L=M&eޭOxnNTKLLUai1u(lLź|I#WLA4wf^nGy7_>jUu.b6t"ҹwF qSz; b&4pNv*E*gp޴E^ 3??m1zK;w/TqV&QWn7/.eqƫ2Je] Zsg*j R7UUN0Yc92YŁ~91 fqio,x@ϼXޭvtn>-yƃݐ+>{s$H癤fe>7OAj% cuE)Bg8xI=E%A꫉ a380E9Za`mϵ2o%)E08 [+JKatmZA-~Nţc?]^>㋮q]|Ę]M{%Uz-d{kv`|ppS^s B9Q?{WF k?%hl~䮒]%{.nȲiRzS%/Ai$nYIt7<ݝR6 np4&ƻBo/vgW!{vGgXQ-{y'o:]|sDsosj79S=8UiK4Ë9o&$ 7\5\E2Ũ,LB"wD(- @1#!$'NgqksVA\-Żk$sͿ{7mU<ޡ'L;Mx\?ǚOj2i^g>vgfϳۿª gK/?gm*͓kxQ{6z HJGHD*fGV g <)/R~^Ig$ֻP?E*H#`:M `S:j>m P mmg@HGt>,D+># 9J=L(B 2Gan졻[zHp۟/ZѳU|rLּV- z :#l 1ՁS"ȍ Y{^Yȸ4Ψ&w6sY fV ]E(']4GzSaT՜ G $EIMVgtɔgFJ_dQ35!{ߖqNO,NQO??/>fXlV~湰P6<yxŃ.TdKr (CVJ>hMd)IFF5Z%/XcL,ws%nC~ mNo,}zX2ƮbWuU*㞾%{=L\äi NT[ 2HޖT2tɣixWpt=[]y^v1C1 'H :"2wV<9pxG*w`ZN%`+=dzzX/Rd9VǨ##P@%YdC Km *ˢT.7A8zZlCG P>ZGvikB6K-x5ՠ7VƵˌH!9ό +(EBS4(^nn|[]GOW^wn'y2O)_8_䫿quG|JJȼ_)1dLƆ ?jݰ^ Le8}@HmL4[|,e-iri%Nĩ2tԵhC*W*Iut9WX3&X UM۾T&V鷫Kz{ ^̱xf3RM^_7"gu/WWV|MC&e yvXǿ-NDe&Li?.YPϋ9oz+j5ˋ\_>9AcA)$U""{)K4M"TYQb`8E5 |َ=2"l1qׇօ\R:)`dG O5qzUs9tta80|3 9CP.nW 9{]Ў(5vKՙR;; ɚF#>+j9j@90z #J8I/%9ZeJTMV/uZX;hS`R??9elT#Xgc̈́4IH)9rlE6f6Atl%d4޷dW,'.YL ȍ 8 .{cXZn1G3=yԏ}MxxgUQWң*;Pޙl!Ei0Q䋊m_Tz9'+jXR`R HQJ[*aff ]c_F_xĢOo.C+noB ;86t*]=SԖ(DYN TP J)hM>"XWщ,%^A[bg#)WҚB"HAWdTot^Kr9>t@pR+nT弰[U奙̴KoWv1,U9д'D=LAx0!aB)IU8K:P[0 -BHrUӧg)y(bK(Qlt6A!Vg%߳J)EQJ+dE& &kzQAyf DAu"-bЍ8YoW3[1ߡG#A[ ^jxke*YX09 䬃 I4oDz4&ː~hH(^%mTd%c6AkR1Ҽ :PlNbZ͠Vq']l+&o*02@$$%Gk]MR.Y|O88AuhG6O[܄v JL*:ETCs!yJD ZyM<ڮNo`rꛩȱgߚMxHPuiPObX^J1׳BRrv4vq U}k=~Zԍ5v8!׍jszt+q&$d'"H6v2N2o&?v~wmH*BnZ!\&{ ng0bllvbĖpZVQ[E 2$&"UȓG+M_9]^?"X1&]/W?/J)6:L.wx}aj7RtW@et H~3'r6GOxtfy*_ d%[,{%n,=N7N>̺~~/|=G[Rj?#%Ji]#eo©68kk2I($8ɂLNk ) 0N0.$ZU`.PvR[>%Z! 8rQo BCkPYX{#Gg Mszxc N:u}b )c뺎QDRx'q.N$PBsl Nx񭾎/4fA9?Ch\o F0:1JtQ FF˥uREg GD8F+g!nx;IZ󫋋:X?{e 5)Gխv,Ĩ!ؖK E:WOgJT(usڃ(JcW袀YrJ(L m!xm@:@EWъV3"k8?|ĭ ah_| ; R1^U?}@gZK&OE\b&73:gJ+o~yBL[-Uчh0,,NipfKY]tL#c)D%1شC5b#7<޼^f$sjswrNd~K'}];dzݞ^PH8y|}ƽ>\ך欴l6O-e5#ͫ'mZ}z82_<ꅧٗbS[8uo.pvkӷ~iD2+?mv^I_re`e,lw84w"W7yU`߾꺑ԗ6W])#taWH_}@~uڸz̧|r<KƗj[~l$OgOz~뿿|~_O_|#:;p"3!8k:_Dqo̍hIr'rW*!m珗m6 ]QhfH<7}OA~ܫ$$`.v2Z-ZO="sme>[]za m><Q60/ǎ鄦:h:uhaQu?ZN"ﲻS|Tg¯U54B 'tg?ƑxnClb  "*I* SxjO 9ZҫV%MZ*94(˵NvՈ.%hՉrM5iji)g@h5}z' I]he4+%t)Ԧ`K^Cƣ-*ah_Zw28mmkmTѴhrV`1tB9A%fCȟ*h&w&hg[8wKۮcyZȳ Dh1$ tU`R$h6%G(jݟ_77ͯsY;Zm;Yh9[3uaD s2{3u뇲y{ psQuw9aBkA+F5 Z*iU2CȽo%ER].t-;~iFhqp n|qx%t˔o%Tb^Hݱ(9năM u̡WHEZl+(#ZS4)|FQ3 T#H]ՇpSݎUz'/n~NnWɱZG'c:-2­OW3lUL!}c|h(n;|UNsytOE%uE%evUtFUtE%mwfr+$y[-[E*暴Cag̝ 71ew}v;veO~ֆPg8|AQ,_e(sfv>Ie^^7x?c:}cVRF\_tv[O˭>wҵ:3?L߿ǟV.f?) {ӓG"hi\!Dhwq` ytc>/.ҺK4V)(k޴,Z]^.lzk1mQւWɁꠇ,Lxw$I\v5A]AIF e,+ȓs}Bb<9+ "ɑJ Փn<9S&, a`{꠩arO>H-pq\ S J Ik[ HѥԞj[oJtq+ W@i\\Z;zjR*U3H]9 ՜+R 6w\J*F+|B@A"Z+kY"N掫V#ӃrpEr*WVARyqEYW$bpEr]1BU"U\WF tABbpEr(WTɪjh  6vm)""{T11& '\X~ vn0{>qtys]sXe]kFϾhr멓;:m837x)]8!xP"2 V9$WEOjM{^xn{>z9e8}D0 R+ى"@Ww}\~gA"+X)Z1s]q5B\$ +U/ז+RX"U\WB;EAB^+WVqE*%T\WpA0'50L)Wq*3W#ĕ삪 \`D1"vR /JQLj+Z֮h݂bpErA+R+_UV\WXJ$+kDe^eqe-+^߱pEre1;VC"]W15zRb⌋3)J3g$v6s'51SOJyf. RI |"OebS4lPF[b= feA 漜([:9{rR}7ѫ)ӗ"Sj\8qָaj剂-+UquGϭ$\`%d1"ʖ+RkšrVq5B\CA" բ\Zcs̭TՃJ8C:a)W$ |8Hq%$&=uAj STnJΦ$\`QuEr.WVeo] T\= *Ƀk^Ʌb֮HWRˊ2 $\`]HZ;H1Rw?ϔ=cz 8%H͐vAٿam9N0GX I-~GTBuGi0vkG4zRjS'} 'zʘn0Bx&Wx~Z}&T S󮏞s  S Pa:w\J*F+8/W$bpEr-WV0;H%W#ĕ!3W$XbpuZWJr-W(XrA \.=WRꊫJeLAB XWVqE*R(qA8S3 8$WNjWBqe@  P HRpEju;2W+ XQP5\S>"PwLj+7s>cfKx3 Tk:gyd@ح k(@˰vI]Lg}0b\c3<3Z >قjw zʂճ2%0P$OKQ'Gr(œ#&)RP==97S&G! |ar݉ R{&p5L%,+SquGυ  `Q P-:MT!@r \`UHRpEjOUeJ[1JHa-W(rpuARYq5J\Ii+D`E1"\Z;HpW#ĕRW(X@0vRkt"z\i.W(! ` W$c]Z=H#RF+l+e)"+R)j jlYj'+RdB%+ƃ+i?3Θ:+(; h7eGdSyo:\Xe@Ԝ*CM9{^$QDj]E@i!O ar S+N S)3ss\ي>z57 \(s,\r])"J䎫gzH_!?YE`p3`ߋ-èOXJ##wG5!WR>ÙʠgHWcjuh:vJ0ҕvLRW̿#{ڧr2AWϐĉ1k+Z:UG (AWϐ`/95Yjp_ ]cstQAWϑL[әAz=KZ誣5GOWXjt(WDWy=kWZ͙R{Gizt}_]]]uWCWo6.hNNt~3@:hϓw 5"B-8k/9[1>O8Zsk<(?4*?Akya#VHMUҢc Y]UZ,lhQ҇p/||觳ݛ?\7ݾ}?Slrm a7E?w;E{ѱ[&&2V$6*0IֶM,߽~UHz7>M=z6hֳj(wOuPXDYDf~z: y>ah0Ȟ 0CxU,롫UGT> =lcoBWVWDW,J:\-k; !] Qp"b\ըcˠgHWWEW:ܧ; 孊`+KVwI\:`b n-t^]u ztȳ^ꀝY ]u~5kW@ v]=CLLk6롫׮:Zw`G9 >Kꆨ[zd-ސҩR/v^9&;k:/kuF;ZNEi̭UyYxdj~~%jk=@i^ZMOs;Qq9LϟWێ:Uѕ'++/CW0?]ysd]?zmO15%Nf6?tv~[ &ғ> O^ׯ__Eu:7?׷?_iuѿO[Țx3w~n6m2 mM#/0{մ}}Yo7JH<7+liGZ_[Ք[uþw=_G̔=q\$~D(~WykFVV_?|XXQ n8 E 6LEw&CE%I9n,E{Vȁ%d~CBYU (ĩ$Ni3UW6!;G~ֈj{QQA}o52Kէ@NUDJhfγ pPBi$Y, 펐UP:Q>&|[4q^Ƞ֙pO䋾~9FńB5 HNW%TGK;L'fBL(y}gM>p|7Y~i[2f=f34ڦ 3}5BW /91*}t,}U2z$9Hmi>jZ. p1%Hv9̓5ŗ, %tB\X+(jIR|$2L"ȼ*c|pZ|RR+g0VhY=ŋ`C%!Y=G=ԭνx3$nCf6⭻s*Yȩ Տ>^ՀXż3TE9¶,9 겚ABH}vs͵ y'shO}Of__b=^!EDBdS0m` /sP6D"E݅ZB  Z451E`JY$\hB;v, ::CxdWXf4 `Rꆠeג&WF%FH!(Ʀr"%5Z/h =]DrE+52˱ on TJƦz+T{)p͂ h[! 0U>oG^sR!Sj1=6̓umW٦ ~7i˳\+hoA.nneFs acJ``ߑr2HuTtk̪k>j98բFCo bPN#z~7?&yQZTk80)QC^"`$6%TQer#`bs"E'%\`LvP R!!KW,֠[6bt6ـ '~V$bYJBi,ve"wQ^01ZKL R,ʨ1@jIU ^,YL 3A. l%ҥKJ*8^'МڙZgD2HkѬU*Si J>{V3(ɐ+*M&!Ts'kQ?u{jF_v ʷҔ/~c@RBm;i*C *mP zx,AP+ 0hYtAia/ZDiq#&08=zMڃPCXPRₑUWh yxo1KLYRiԇ sȎhVCMel$Ct.Xz餀+9b4e*?"t+NH1 ՠԫ_wm{s-nnwW^]G P 6lqIMF=c3*D[oN_~Bq/WnܦW"lt/\?j6u68ӻ+Iqu6 Y~EO5n܏Bロ8={z۽<}{Ie1]=/oc8kn(gC[ӥM%ڭN>3qjt .߉ =tu+%T}B2W'z.ց( 'stcBNNNNNNNNNNNNNNNNNNNNNNNzN Kl횜@@v5N |N j8yp 'p 'p 'p 'p 'p 'p 'p 'p 'p 'p 'p 'p 'p 'p 'p 'p 'p 'p 'p 'p 'p 'p 'uyC> Qz@8(oA8h8+mֲ`y)b &xxICFVZVIr/ ުE.U1;HX{YY @Y @Y @Y @Y @Y @Y @Y @Y @Y @Y @Y @Y @Y @Y @Y @Y @Y @Y @Y @Y @Y @Y l@"WZG p`] (zJ j-SY @Y @Y @Y @Y @Y @Y @Y @Y @Y @Y @Y @Y @Y @Y @Y @Y @Y @Y @Y @Y @Y @Y @G t3fؑ*^4TS|&>n&w /süͻ/|}5Y|Qpp$\GppQM %cYK7Aʺ^ •z(th,᙮#]Y.Īg`f1㺞ڷO_-c2.4;K.>i2U'YѪj=)cQt!V*5tBؒPBY半?/u nּעw}V:)r #6@]_?9DbgE] /N_`U_m).;)Y7Q(B]qv9Z#7bkZ@[y7~oP ])B_Mۖ;6Fqtn `kuCmH[T) q8Ҧx@%5t(GJAhv֞&QD>׊\d:RB\ezt%9% Ճ+K`COWHW3kQqN ]!\MBW_~<]!JHWK eYB\;]!J!2]=C20bDWX ǻBYjG PrHW$΍םym-իΔKʃ͞_o$7 LDI~F6^QE[4@ksȒҒʂ3Ϩ:H~s3mq1b&7c_{7^Mxl3/t̻[y˷a'|1 rc)q+ TDt]a!jyxn8SNj#G2_7~뒝֥sw/:z6 ZR_UebPZ-EaNF;'9S/4e ,?*ۮ*H^H[7+}@W(p32RJ-+YYf Vh>|OνP9L`eLVjMq Z_&RrΫR8JB*qegB0SnTN$g:H2}6GrHJ0Oy\ßBCI,Ǔ@W4C a8]!`CWW},PZUeg  ]!~ t(-t "s@t ]!\iBWVc+DYҕT05 BZQC+@+ ;vB"]=Gk9$B$0vDK~ QvJIِΛAԂК 2]}X1$ 1B M'%!]ňқALp+P ъϻB2/?GD*Xg Q j \oLSآlLT4O6J(M頕d}}nlc357Ǡ:xhg:C?eSˊ~}M~>DjV' yY9hZhu)e%L*+!6GԼ?`v/a~s=3R8n X}<]n TƣcVpv\?}#| ^*<|>X^>ϭpcgL>jݲ׏vpO<-v۟]Ƹ2M]qV6.QY .aRekQ$q K 1 LSFOz6}t1*!2ZVKY"F' @]J銩`6]`GKx `syv]|> YNk7lz5d/vب=gيӺ.nq|aZZ&MCຣzhfz쵻k~} W>6f{ MͯUW]OMj[7ȝN=+]5OAxg%nKmCsNT*nv:U[JPujw5ﭚׅﯼ yn{WױGk>w~6<喝p~?Q5m^ƿpS6/|]-)m=8ڳC7BGrR2;騄R첧&~/N߽\A>S'nՖڜ}F^tsBϭ'jqS63^>|Qt̛aBzv,G= 6WK{l^lSqM3 H]qLJ)(|T uԟi.RkDj]Dn(&Mԍzc"uYj,3%qWQ]Q)b%'gb 0Bnn_.};O_\cuwp=};"\n&o<-|;fW&?1*0NhPZ9DA$!fh%L%BiP9E}Gv~A/A8B ]_ƹPS CiRFGU,.ъGyiycg[2~s@6d4MoJg'/PqTdN,14JK%.ZaxU|;_0zqݸr J}Y0Cg4VкEs'%UBg[.^s_~qM+0#^uh_xqGn9Fֳ}@/I.^U3k*7vI 7CA03a|fA]/ZΫ? Yhag~h}]߬MW72g Ne?=\wy77ydE^Jl.ب8ͧ> U]eg8YNpr b{FN(|9qkvKtolYN-Y'ܾzVUY.E}0l,쮲WFD/Yf#a_` !ơID(m&auL{h9񮊷%4ͼ\nni[1ʺ -)7ʁ)̋N6$mM\&?'YP37G]OߧtD'{e,x.&t΄j2Z>#xt39=_I0vlpwD|dH>g)*@{xP*2aMP7 [3O|݇=B*~b6Dkymk;]-g4:?=Q?!E$ISH%ąH%޽4Y H:rkT"LIUr!>?")Zπx_8b5.%LZ*Q*kTJeW8 } [`~U2鍀5-ٽ@ˑDpgnZ ˹?\brҔ"IQZ]0U( ٵ{}qzl}qv<s?#lU@DEYPTSj"I&Rr.hό:DNC@q$eޘPܿ~C&ʋT H%4Nsɽ1I_ RK`'"`0~d8,)c(*$NcU jop|qvswcxڛ i9 j[^Y:Ou&&aZ /+C%Ӡ"7B@J'}J*G) X=22:2&$NPOC .p:ǥTb}3copg|޸7x(Pe.|t*;qL?bpUH2Au̇~x41}_PF4=3Ɍq羓4!q;J) <>##uW = 8%ep}(TS+c$53dtU9TzKI3umzg9cH h$"ghKEs#,i%Ne, n%b5rG^u 8]T6wYml+]ezv|LhK;NBd@ jl D8sK.2 }/w!"-`uZ@UM0ےǧ~tvُ*!uq8&oM8`|GKĆsj$ơ "i;Kw:>vnNVw 3NѠ Lkbr1$;@Jǹe ܖP'b&3 zPz2)mE ksQ:\R˝-}g' ̶r?p%(Xϊd{l[J2zu3+Y1=^WxJ;+̾q V-ˬBFzGص)m|R>b7_]񾾅lc k9LĸYd$пbp5(0ʘF7:e3ͮoǓH(:mnCZ; f?"yәƟo:9=fwI HDR "*eS vFcn۩MѶe3"`=o˨y. J\+dF":)hEDTg>%=8)uQo B )g8ax5rYɼr1}#֕NpalHГ'1m(X;"y!|;h5$IJ*v\f&e&L.y s~UD2d3[EMz3fU1͜ -EkY69_ nxY,I:'JUҁ+A'3A;>q'>>3.Œc9t?Fʢe Ij"t%jM\0T`R$19V"&j-ʅ\KIOq3W9V#I9?p|\J9oٻֱ^Ɵͮto>BxR!o;cFbiHhlIr'Y8v&tˊirpmt\e2qܓ69.2Ȳ`sT9Ms7߈O㳼)< mYf,ȲTVeI J)N*5Vehq |9}`z#}sU&)<."2'ER&F/dF%1B:)s2 t غf#%OWZMB$ e$ PJ6K?ֆȹL]JCUҐH 3Q#= C N"ahUnE)~UzaP1sL-89fE 9 "Q(+-*3A6_sqn3$=?l),"1n=%ۃYT%t-HޕyU̳1N9q+˜8 Kp`1  6%C:$9䳖~lnmNw%=hg/|"3' [L1]I1 dgZ8N"= dO].!x+ MIn`K⠬ìs%rDm:V5ώ{;eYzKߟ}0j-feT:#h)LeeA-zeѫk36^e<Ma@A oI|+M:t=d3I6{EU:wtBȆgOj`/g~j,Xl>}|2΄KZ`wY^ė;d{J\Kn|!ͺ$t2Dry2|ۿzÝԻiO%5M/xR`h]c} w.}ZrdA&5<2Y/D0.$Gd/ϐHJy|{=ER ox-kI|_ׇ RDz9ҕA;)dgI/)Hˤ^H+{4|J_9\;#%rƂi-^f뤊 qDI U8H wr`'Z~u4J3wۻY5<(4Kd[XQCȍ%Rr9'Iߐk YԉQZ%JVQ@,9%& ¹ "mmHuiE&fUe P} gu a/WBԽ 6:NqX4=%cA8> Gx`|hh$$?繫:\Tazc0kuWLؒy\ۛL~ƛ$4zZ8r[2d읲C"4/{3'{5sfKY9:z|bNe"36m $ƃyۧ謐Ϲ|:{L8Qhӗo)ѓ}m˗^{8)M ⽮%ӭ9+-7[IW IK(i> i~`vyЯٙgGoVubS+hl|+jiBohpdȞ6WN6-Y,1ƭ'XbŃ^9<rs$!7 U7iՍZri)~O ) +>Ob)ɯ<(}Yu59v|=ٗ GӃ8<=_7߽|~y{zsk:'pbٵ$ f^- K?5K +ﱴw=෮+P7+øԭ- a#n85_d+AԚDmMl~2 8^JF^RO!0v@`S%n5|ViM|;$I]2$A tb ,Z"sQbTSDy,-滚h+(  |,H't6demUVQK8uu:_mOZ[wB*7w|rJZמWOjɔ2a>BeTaɸ[/[O}Iұrl$:4juR8_v/>yvSzvķq m.]l*N)}$&Ǚ #Gt[yLMPz^{k?FGlAK5( /#@⦛yC'Wn7|KrCg/(F׀ԭTO8C**/KX |>>35 G5D#Zo\.a!X?GoWsT?Gخ>NSA;bAdP&,2RX䁗F1Ci}mȵnȹFY]r=42}yڔ :pzzA\g:!AL.鞫CQpYAc?J&; v*7%٩|5Adm,{*\0][=Эh@cܚSq#2Ч] -"Y4`a\.v8ఁ⋬x,i5R{fdI[$4ۉ%{4Sͮ*>Ƭ%dMʎ9$EL6e#8Coy9z8rն1BJ ȐӖ}UPN-QO)F$@;vo06qYNc!־Ꚇ{)^Υ_ i9,mLCc:_Ʒd ( @d U(xS J=2{2e@XȶnIɇm5V0)BTGzC@ p T]]v9^B۴> 3xI^=`6nO8=.3 {6luYʉhSF,k2$q6k9ǃo5K tɁxO "S8B9w]^zwe鎜 7F9IeJVu<4vh|׫3?OQ볳b|_̾rEWW~+ͯvg⍳sS~k3u+1IWb+vsU2LkLWB~j>1?*EArr4x{Jji巡~؃6 @z(%#Ձmevws]A}JëwcXӂ/?ϯ/%Ų; ov ss5ׂy 9d,jTX|͚x!rϲݪma ,۟.u׎$ov{ Ηoۗfx1 \$e7U?s{[=dr25Їc BuhcIFOK_t[/WcuM˓˷ݤ|eq9c>?~a 'i)yP[!- ἬQܽaI)X0n겂QFI 8!/4/ u}X/pbddlzH9U(I!c.ӵjdBΚț'k}q_̺Y9 t|˯m 3mwa1OPePc^B8"{(I(5jq2A 5p[yNHj6 c,蠔4dTTJֶcE8&|MUWoB')%mrT_J, IR26SdtVCU9-Zu gpQs=mRN%l.r'dMN<7$2T &/+vgvXpU6]TpD@ JF=*.S.NY7)c{Esv3@-{ǣ ,mC[+S~#3?\":}g3H䇣_@V <>CްORw{X4xշ\<\]ϿYIT;LI)b5fb׊u53:5ʬڊ#hS%>cT+F_q9$ޙ9Un\:B7sڵ6d| xV"7C?'_zNښʙ)VX]0հDhP[֙l±,{ T>Af0Ҧq>9dCȾ S*`(\M^ܝ _[{yC'{+gSE WTBX Q5D{(>1:-ds:&[.U|L! 2(,:%䢒" p%"jJBs>fp$f;-1n ͈9#ҔxJa !(.m K$n-l v72I/:W}9} (ܸX4:T<Ù(N{S(%aԼ :gnܝ?+rYbX6uv9/)/Ny6gVhem!((Xk(1 MAK2KI S^D^ܖ|;WRnkq8)9|u'vEpSsU?#DL`/GĤQDL]ĤYi$b5ؗ7"_A{ pƩ$}<:=ʳIeCr "8sS-+8 g.φT߯٤2h%rkĮO%hmg5>H2r!dMr(Yfjkl%1pƈVض"מH+43;S{^ܣ2/oH|8ZS s^<}|ޗ垾MxWK f}Gf"kT@. "?wjۚ 9WlV02AA%Lm[͖mS֩mGUfR*`9Q]]eGŘn=77o]k͜{#M}\>]2WR =ўk: ië|ݛh$Rϴ $ALsܓ[C׫_ZC #Bo=ph7اVha $/P20էzלr gӄ1q/̺i*A;pkbC4c.P_sɗl7Hh$ة$w'΃!K:JD,O4&RQƀƹRӓmp8Yxk"h( ! ͜* yi*;W3 c/LwʹA&cP2$T%mɹt\``"`Qk]"*+hr(ȡsܵPq3a sF s2?y;ZP $`!\YgLE(JJ4Е!Yc:,2%gYf@ C[w1sCIZ%Z* n7]9AП&}4YҊA|-3Q !hsMy[5EtHE7D:Gs^o{.{v=u|m{/nrk<aѭeWks##Fљ ';Ts iztl8W[GBʔm2;_ό+ZwhGUԁWiWCPuUrPpQ<]3,ʍDTLDG%&O䦕&o~ JIa-1˔7>|NQ[[eVT`ƹlȱ5r_ h!䥢|Կq뗽AZz!Vbr2m3 |[o{obP-s̑Xx)jEDڟ*]SI22mn6X&gU+kw$2I3fO Y6LNk A GLk;U^kkq7ǣpΕ,}f{{NQN|$j7լã-9,2zPC132#9wbW^UH@H!V2c"u{! cK +.de,0-4xʑΥY)ÐNόTѠKq5Ád[ $~8T&EUԊ:MZ2$Gp0XBL JRǼ 1E D~tr@b\GhdN!H寕(hG{,uʜ!FkjABbuIg %Hۃ7K)2Fg粅 [l!|Tޛ {1fCR)mL TT kix)s*x"~:ow߮A[ƠH@Uv?rደ0 Q,_ ]bgiO_xV|lzY|qwO~Igї, ~;UW}?0jڷI (FӄL1AQz>59 n$(kɤAy`EIȘʊĭ4NQʚјvsq-P&}0.R`Z @-QB.3K%b.F:J;Nrp^I)U;Hn*;ԪB8_v}]cV2}8TuĮ(W)bȕKʁrB@UcǼ0LFbgoS_7q }aPR|ۙcSrSN}w<)BV#*AViyalqc WTRXY ƀdA:9ѵ JjݾŠfEj5s;[%v$As}~ZA'ZKr-Vo'wm)4ݧػ:IބڕLuz4ɵgW=N>}i04+twyl}0>_I"^ݘW1R:m7qhn M?P66nzHI~8 U>Hq;-NӉn#^*+ITfOVk3iubv)L1{]E,AdiHEKnf2R:t OiżN0@i;IjIBhC~qUNim:~֥C ZSy|t [AubnTn*-mԊfG{܆g\tJVF{6Q&Pa }h^3#z/Z2nta.#7J[T^fK!QVgZCC Zѥ*jz+Jn=Uv6?yy<i׭Xء9w6IZΕTŶ5 9@F J-xIFXtKO*zm/_AWQjk0v*#0$ؕ 3gІ;|@HQUJVW2j^yir1rީ,]Xj?27)hCI]8S+L}'KzٔW>嘐|8<<.ݭﭖn{z3E\9cYpA%ӌ BUp&IBp/GO1ӶL^$rO;g8N4N|sǐ!-GJg):c hUW, 6˜L$O[ܴN˺?^Fn GODͼ 3VN8B"9<]yhZ>ﮱ%orCV语`W}xߑ'5oD :i7Kfp2byE-/{.fZ4"o3Im~2c6 +l2|&lMܧl|ՆBWU󸸼g)7dcgz:ԭ I[ {U}Ξ[' ϔKA\u';̮aE?v^8|XS#o~+O /O8]D$w.N]nRovv^]FX(5ep,nZ-947 G'Mѫ6/ix!+ EDAmx. mWlyn&y5-{wݵ7q~l.՟~S+'w ;j 3d,s5T\lQS+'YeHrcSku2~B\-ѿ^zDks@pW\2pT\D)LW\M2WJzp @LW}@9Bj}L^ \5z1:OeNNFo|u|CpGA/?}5MͣgssZ95Oo?n%].Mo7}v^NkdM y4[>i`>jȰ>VR*^T84.E̳ ~G ۃRq] QՁk|hyV\__)_͑LQTJxb*'y*kJU.JjeSBD*ι526*u<Bﶲ +ߕCVl;Y%8 RTY VfOH0s)K.+XrZine*`}=pkpp8{fzOq\=Zx&4_pW|mn2B \er<*;\e* pі8>u(pŌ$D2k^^\\zӳn鵬ragNMQu\.sem^:8CDдrRdi(m&1yo>i\Vb=6$t5O~| ]W1-#Dd Fs#jy 7vaɈFIiSŒT)E):NOIolfTqS[V|1:sGY(-뜩J V&ؐ ڼ͛y7]mfl_ֶ|3 w:R8A>es/F)r_)&Yfi!Ⱥ@7ѨD2XグB (+2OSpIKTzW]|Xf>x|_N|CKIpRMN G>B_U\'WJI ʠޡn谿muYzt;im:>)n3K=׾;QmO `ZZ0;c"x'PHGu:pckz,]5FV6—Lv]5s5dwД9~=4Nr/ooIG=L]_g;۶jVzlys<.V5֡Wݱ6Uf;hZϤiiiޛMhZ&1݃Ui]2C:]FJzUnH#w&QSR$@B&0ZetYɻ'Sߛ?_2wvTQ4Kz7mhT ;7,0l,J ,rJ$)*""HXHӾ s֟#h)$,BQE5V2)ץ,Z$hj4,z$c1*p@M ]9C%BQѤ69k5)*|$\ͪ2WBvN[m6Qt2Kt.] 9{1Io NEƵsJqq"ц\#(7GO!Z橏¡ht Nk#aAgP-ήPK)b|r' JڞMw0E.'| T|M;nߏsT~:hIhL+B(9BxKZ}J5)E ׳XBrdS},!ǂRVGԸXR qƶX cpWY3ܬ݀ݯ_b<ʎp|j^$ES%5>D3H` (&mN8 sd,KRiT!hE{5bM+Q?R8"FBx40#aiAb9\ K w!P qIQ-FԀwDj(S>%T (xpq<,[!'e}ߣpӑ|5_ď7 D,O"n( ~*߆S-8nVdR &_b6Aa>ZjF[ tv`&Q \ r /P E;in~QۭvէB^GT/ѐc(6q&9R F7ʆjQ]Q7B&BD‚6*QfE t`9$g!Kϣ592Y,׌3eEFD KQkC8BK(ĨQvkxa9+FΆrVտ1[ZQSF[2rxnO$u(ށ[PM I3ˢEm/ @)CvFzGUN줕©$,<,OZT#n4\qrZr" t"1ˋ-=%?od#1PesP8rb\sN-j6]#]"…Mb,·<Ċ;NRѤ@,w"D>744g_!*^F c_1V: #%:A/ycvܝT#29RΐZ'ˀ!䙂M|#eW 㛿 @rܚJ`*V;o(R'e+Z^D`F'Q$D.6kܠA,1FK )ZrK`9H#  Z,8eT!2i4nd.h,J(FΆY+4퓋U؃qDzxMQLS.6$mMwqH4%Àssp X`CSx4YQݯFzF"l*X+q_,n 6&y -6ICh+uPd3m/W5P1V F :kåXMJqDrI ,@ <ʁ{A}YPdyyjSu2{f]c;7>H:,Qx6!F)9AAڋ=+!:YrT[ӝRJxV90iTh@EK!dQ: Z "-LH2Cl~[`Dp@BH՜2G8: i\kF;"Ļ'uSu􏕥8(%jW|0shxWJ]Vg$Y}עF|y8;=;~.E^%|Ԋ\լիMҿi8)|'GWPY9RiѺ7/5gci~Mscb,k>d>^5 -糦07ZM0^I6DnINۆa @0[bĘFFF Y,;x4\Nj`HB l˗l]"P3D%ϷI(]OI~XpӏX-*Zw6oWapvBO?7~M|󛣯y{Go|EΟhf)-$X&G #hy݇֬i; m˷^Lq?|qxk([]_׃i(/ͩG9l5I+U(VqMl~:O4VaECշ7# !0À@17>G|v&'N$ɿp= v YX W"`yR1,"fM24BDȽÈ~f+Zߢ J  VRZGye0cP8:kgk3q}}'s㥋Xջ<|7;ӅFGNeQk/,Kމwx:"1 01YG/u W̤Ո.%wf"A`;'4OJ R0G:$P*D)eeJjL1ymlr 'Lj Q_)H:+ cPw< t&F t٨3r [~2.! &C@yQ<r(+Zt[5%NDe{fLܮ{=췭O >I%U@bIkUNF6 J H{ %ÿN߅jRhMfbl31] <0V+jݼṔ 5,w`!\[gLM(JJ4W`|HywZgo7@ C[w1sCIZ%Z* 6]9AП&}4YҊAR@STD E5A tF@ڧ c3o 5l'.X}϶lknqh}I^|&]L?iSk:eNzK>Ztn\F޻5kE4X"ZӶPKkvRښqzMuGiu-ϯSUM{6VIu>jł^Vlfꆒݗqj-Jma*^kםW,.>kS˖qolh'89=VbGkR9srق 4+h]]57MWaGA$* L|/zIzO)'|g/h~z16Qz]yo3Mŀ.}O|r89?H2{p~Qn]BQiAJ3⋆2QVF| $>k65rVݫk苗sx/ JJ&YH6q:noo;߰'y5-YxBmClXHzY0<1y/w~Fjy:kn3h:/iUs8BUc*j{32\q*Z|"AIߏVѹ][wOǨ =$ՠ6ևÜvȤϔOGy>;WN#WNl>"@:?Z٘J6CFb`MN 7\ N6:Y`Zh*XVh!rcJ0䁦3#B4RY.炝s<<a_/<[HYEYjs"Ie|ɩ `$6Aң&0ĜChSCSS_!FSR`JHǽ H:keCezB S!12مك7tsBޅk-v)|0_k/ǘ H)FqL TT kiE$sNy"~WuʫO=kʦԟ^<g۴vደt`$tyYcRJ٣:ez=\d : @d/Bn0 ,!423}Pv'A7>d0S+m}zJ+l+%Д@?,=7<9Y(&Q}m2LN^gd1 [y AYPt3pVv3*+4NQX*vSH2q2޳45v [ 竀īƧO d|`[x9PST\"W.)2&zm3o'̻@oI4זM.k5k(v3}go;oh*v^{qRY&g.3$`eɂt9s`WaQ5!!)I^SY"6tVe zS-ͯIbhFT E$+2-U4Mr#ٛGm~m>z?_Z't}/rP|Sn#NW ڹCmk@jVCgT9JW2ΧX 蝴)Y]˨yɵI!GKK;L= y -;.,f6mD'AdT91UJ,/VXKIaÌ7ܦ~hu"Gr@Oi]W팜;mI »)`s 9IJ3ɐH,e\2 5eh!XC+oN6Hd%c~0 w#eڞgtJ괉U=NܠJ$ !=X6Uଉc-Wώ'.#bbPKK&Z,5 |_G!taGQ=hg ?wbZ0AAkj:%0 !|7oSǮ~i ;JR/81eGdB 5z_Ats@ɒZLGA rٍd{Bz@ȆPi"XKg ]pAڰ4߃mS)ܧ 3Z"}|FtZT;brL<0rO6Gw=T4ᴮA^xl@m[KhؒuKKBּw(nO;}h8Pc1Fѩ{v)>Ct :Gyg_ƅ$0h8fK@3lI0!/5Z4^fY?d )j'/ŬߦZNg~&n1h֑X_CM*l+}~xB gY$zʘ@S]@ȷś͜fsvǢ^t\NC,4Cgՠ9)%?,gׄz 94/ _״ \Sߋ۹LXԘXK|iE*InguG"y[W-^>/s׳rp(h]>wp1u/Y0;I GD߭Jdֶg}hj}_߇fIޟ n׼g5ͮB⿷mD'{{M|bԓ'<WI|z_8p=?=ՁOY@{E0Ρc;۫O>_܆GlMJ%ikLeKw [lWu[5pVKTqeʢj_D6}ysN`F v<:_~}`̅o[ʪʁ/榇J;4MTZUjq3^T-8aѓwQ E>~;'+ ÷áE0]^w㌇B ۅѻ*i;%?ci\lbWWbihpxM5הM7eGݿU*M~gKq/;X'néysx}h~ZuxqN7K/:2ѷXw?ل }1Ç_\ ]-@ ?7h ա><ysY^Gܰ~I״;u({ƸuF-Gxf~Xu5+ۢ{JQvgEb ,2cG.G|7LZҕa D:B\Y*!bRʭ*6j<̲*~]!0HWTtEĮ+tYWcԕcZVv{unҧ$<KO:]ǷO!vbeh̔?W~Yg&9j|€al*U:)ܞ}xMU\d4Mʤi}.&J Yc47ЛE?շbrf,WP;BmOK1jv;]]/Z_ (orM~;UMc^ϻ䫕0* _hy9$|;5*a/=S{Hyb |nN֟aR SQȺ92<42!]dtTtEǮ+ F+3X+,](mn]QWHHWt2"tZWDl1`YW#ԕ\IزdtE.+Ů+38F])JHW,]2"\eS]WDisj`'+6]Kf i uE"ƨ+#ZWP;jvaf]PWVpkUBB՚IpU2cWDg(b ƣ+nj16<&)sSΤ6ܶ;K l *?|} Z;~\_J֓ML/Rϥ}U3QJKů Hכ{`\c\iҸMDz'm]q)Q9+ZS9gLe6*d;sRnrJ4\[r.?KLvQP%j'Rz⌒ZUG()EJ'pu2FACNӓS{f:c8>||* auBkh)"[5NJe] "!]!,]aŮ+Ԑu5B]q)Aۄtkg(](mn]QWR:0ck1|tIz 9װЇ(KW:@W:Ь dtE+.v]el%YWϢ+ZWly2"ܡ ,z]%Ϻv6Y€-$+uɴ1lu% ,X2B\l*"ZbQJ=VW,_7 WTtEŮ+Tu5J]i 銀Ykm*BZǢo]uueeI銀 = wЃhuEYW#ԕٔڭ`B2"\̓AwRuj3Od%0XL;B*'^nٟ%(aǸAXbKe\.]jk6^k&:AϹA9k]F(m*AOJdF?Ǔ(]}9=9gֻ3&@Waj耿0\=0Z;PMelrrYWf= %+ox iP+QnDͺ_&+6< VFߺ"ئg]=JԺ"`j8^WDMu%2[pMEWD+!v]JIc#"\hucWDiDu,t:ku*B'_Xx]cWcԕQyS:VW9OFW+y2HǮ+u5F]YL ttE+uя]!%@\} f9Rl]f]9.!]Ћ\Z e]PW\KYB"`$Ӻ"ZcRꭥ8ƣ+FHX2D*"ZbQJȺz!ʪٱ1E J1zLZd/^O? 6,so.K,c^m TЏQT?{k+Զtvwctݽ%/}y&ٜ`ߺ,9́F Qo쒶9DpɰIru}umǷ= ؇j%|Cs|C76JmFEĢΩnޜ挭Jt*ۼ_j/P$0\apX6gKe֔dL[C_bByh9E3f17Z讪l浯|J*56k޵ql_!y]`0Ȝ08$y$8kLD JgmK˖ԊG;dl6WWZ{ZU;Q椄,@Tc[ZKvJ>,QڎMѵ(%b'ԴIE[RHHH I? ҋHch#JK|S.)qC_DH,&pZjS6z\"FQ(X'餞4/]FT,hCD ɞ5)0rS`w"R+ ]+(I׆R ӌT"yˌwh|z lEE@ +ٽ*ؠdxiЮmtW*A5V-Tbպ*5%,dp-M,@8քBGWX ĭd Hqi FCFzzE{K05n0 `HJƬC6,h  "ٻVLtzRyA1)vx X-V-tC*D4`Ԭ4 3*0 sG  QA\R}2ɮ3¿Jt"T]!z@e gS d,d*zbJe5͐jPoB+"X2nP(SP|k(P,Lh=4v#]7cEfJWukʃ AGb̂Gܤ ̈́#Bl coSL3H ̚` Ukq*9mL:v_ bUPP9)ta7 U BI[Rl eB+mՕHދ0J]VkD5{RQR@m32-([jkIk ˨DPZM 6A j"b C$CuH*мGwU+czh2& ԙy?_nZ1#.Eu44U ULl^vNR'Dh6`V9;4&^z.9KZ v`c{{]`!0h ƛAy@xc#T Yd :!JrAhRU! 0阊'8$;ZL %tF\8gP4I"i QfxdeZCP`GMtx L}h$kIu 5<o 7tXTu~R% 9ՠhKP-"V1hFy¶e@TD8X!?%?=9~ߛlӮv\GIWa`0Zݣ;V$LC'Kac+i0ۿws'QEEmCYk Q^$zhhPAI;ߌo9̈=Xe{80(QC^":$5\Qr#bns"I%\IbLvP%(H :Ш3) @)w!nkQMƮXB߁yE!8ij)ڥA$\s- ,R7y/nèIP(EeQAR܌EEH,{adzw:H)-и'GJFdUA?FтΨ19MAk.YqFO -jITAlRԞK3e2 b2 D hs@:'/[iN r6N T&7%G 9@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; @( ]c|@.@ VQ@)C):J@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; t@N 3q>N B'f"ΰ' P҂@O dΰ@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; t@N"|rjW,xASM骷uz7~QVW@ Tgd\`ػLh\"ŸDh;uKO^Yn6tAPq)ҕ4dFtE ]BW@k:u")ҕAK3#"n> q.tFS+By]iw3+b>,Ch߻"GUBa}89ں_7> =[|s$ddwI2 aϬ>z-~ᛯЊВ:עg{C/ Yև&MLq6ǡ+fT-<^~;O+C`4J*i ;Y~fυ _`%}V4wCk㗩hӪh*}^*H7#lF9(ݩԎ ҕ̌J) ]\BW֪S+B,+bFt|S+BX]=E2ZH3{WA+fckPj ]Y*!gDW ]};m+6y/+4zFtE|nRj.tP'IW^G "~6tEh?J%S@9m!CWw> 5'dP=dz:ti{?;H5hL3KTrwVT~sƫsa^𫶹\P(b7w 3xCfqNk+Q|BlԐVS9 [ht=@pn{;OT7JźIc!:X2oyX3yH~}}^!oj̯Ӧ>϶\ ߗnZê՟Uw^o|sA@%!oh>|wVSsMݷ|o/ptVٞp͛o0l|=e?zA4\FiWo;nAgtfA񛖥ہ~x zD{8| bKFp.s*˛mC@{12>J#pۓへ|u|_nk`68AIwHoGywmqֻ͡q}K\75%^|-#!.~ya+GU/.xE/|6/'&},˽4*\/w׻_-Wߦô1|ʋu+PےTustNzp^z= _Cڸ$}镩Q9k*+[X){at˺K u88^<z/BWAH^Xƫ/g>w rX/kAۘPPks-v_T6gKu%ޔI9HzWzFi]oɕWiUaE6  d#u8HMYc/nR$6/%JXu|W2BrQ^`"%A֗$+3/TKbz۲F8$(7 $DP&Ceb.MVn]@%몙U>|)1bC`q:W掩n;>䥥VJ\wΩ! )*oz 7t{7FG0-N_ ƠB@CŊP)srPy6γ%|[1&@d91;*'5ޫasg[iCI'a'ԛDmf^&.̇IւqyRYEF .%eYC$% H2f@% 1k y"{&<¦blSQ8qWU3s:̕2wYIUc_#Y|NplM>/D}?ϽA9[{2_A!TH2,!21*"%CT^$*ΌZbZ2DjA>fG7rRmgI:]BlVky-N>YH26V6*]]oAFg DԴ¤HQi>{NrUDRhr\@E:i0FZ*J BpK*`+lD Ȇo^ yos:eH/#ʔ'$թF+*Ljq\Bsj:Ss iLA$#AZBa(Fw\:/ӕvA'=!a8S3sPN ͤH ,$A@B>%V |F{@0e<`M锂Fa&@eiy`rn3kϯ_;/sFH#F*&bH)E&d/1#['i@(ک' X05DdB^JRҨ2>hCBjȇDH{XxY)/|<ȏ[;v{zD\%C}d*(&ѨXr*A$1fKp[ƅǼS聧 LfCO' Ġ#IXEK1# (qH$VrB9 7&zz)Z`Ńhap0rUwd*dQG2M2Dcxc" 41Rꋭ,Xx2YEǚݕ  mj/RTZ.]|v_*y/=~~;{05% lVR)j;oVXL=ㇸ\{ɻOf0@IтOIl$Z>.\J9vGb;}:ͻ{=tCAۼM7&.z7r^z~mNp~M)= 3a?rӏ<Ϝ[vx"3l28Cy/?o܌bv+w=;@;!evS wʊX؃\:>Ѣ#wjވ ⴩!x?`U%blR~MBXq|R} $ !:iX@MVa$o- sný}ǻ {Uö8նY7yī2ˤR0|]7J uDdt8Z O4_߱u.L"yg0cߟr'PMydҀ< Y8.wn$B~]\ j?O&ߢ Ovxm\L\)v&V e: tՉwK8$ e_ 0 ,`Fhxd9ˠ)gۇa{Lbi}}1RhNYC̊) Qv|G F"R(49.\B Zd"i֨DX\_'>1"| pюp_\L!0i!TZ])AF+hD; N^!]T5̯ODN}_cxqNv6C=`6vóq@ /?zUjqO^ +l>=9LdP83\`8$G0WKp'"p1Ӳq Xs!!l ErL@D3AkML(P}ŬxHKa0)~'x]ru=2 E޿ $aZ !hTP1Rubڧ^Jfp3(cd2&&NB1YϢhRN2.2{*PqbXXlf0 vX4̸wp 9:Guw *7MnFz;)SU@9HjC̣=z#9:g2:Ա) Ϯ=32lr^:nGMLXVm #v1q#㊊y2]jvں0j{Qg#\9XLMR-Nʄdu epI*> ԐRZtk" ?RH1H*ՠ+َR ]x2:}q("ˆhzDqc\5#q89s\dQJ#O$!'8c J gP@cNo %Di1HMJS`)㌏>|`B@M|NBEn8p=bLpJbP\q=M`K._&? g$qE?i2y\iORE#j(YXݬwmF_YK{h&oCAq嚠wC40ꨱ%W/9,ǖXXyj9$g~3Cgh"ƒ' @М y}r $ҽ$t϶9*qU+V 21t!n1Mhȷ)D픖:{Id}Bo$b:m'tk9#" !s'X]Y!2DPB ԤS>KћWJC3CZ6M4]Txf=REї(6U!LbΞkLj P( &%;ɭ5= EHeqtL9FBQy^ K eS9s,;a@#'-7  L$@&g1%CFO9+:Po@U`RR+FF[qp%Q<2Y" ,XJg RD.E.ق@N?0+ɚ($828(1kgiڣ _A VgV~R'-w !yp"a@L$""%]N9;%<9G'Sx:{F宮dw&-<5i@S{p.7&*2t&h;p{56Tޙ }Ov,|M8&+ͳT2% F a%P.B2jJ^^<1S:JG˜zIQ?B#˺{pc]61$J" 6y`+u*`i|tcG($4^8[Km[(Nʽhy(d^&5˒/Q+,'xKvRNkg1ܥXuKC^tL!E!ʢe#̝a%|٣ӣ5rnFJJ!BJcY9gV%j+I_,aL~lYaJBsb6C̑*Wx.l6)9?RMG(~cD L6ȥT!'$"9*@+q+ C4Lxtv 5a(Wx􁗈FaȦcI(Jx$]TM ?ۿj(Qԉlt(ir8KV I>k/fi[AϚ_&v*ғ-(3u4I΀0/.šijgῖ+ %PkښY+?v frpi5`'aI?ڏw]y…qY{Tnn.5,4k./6佼Axy7U[f7 ckR}&t d'Gӳa)jrۺfQ"7ˮg+ޖ +SiT4j/kYg{Es³U&1=wlp[aveXmlڔ[| {O%cDYےTtX mfY^ 45E#XFhr1Г6LۭE'׵NʱѴҰ"q4ճ WJ?ը=WzK㥚xea{ Ď'ϟy'z닓>'IF`Bg rkRoA͛6EM[]qC]^˛OR6$ěI(/ͮGlҚaTA0Vq"M3Ta?Jv<6dCXB}&`{tݚ,? $#ِ%2ǢI^hI#,Eb2BT{?k/8, oO,#~v;5[ Ů! Ü:2 ] vhQeG2~_} }'}G.;U}΃/3G (MXC-OYƋ\(LA-ORK-9ə 6lP >I)F_'7ׅ]YAvmYr]3-̕#Ť,vގ>P+͆ R0NkkbT]R^K%Q%#8eHnY8J{-%ާVY62IugUȍ;=Ց)U~,Ŝ: 6]9A?LY> @S AU2)G߱Q9vX_&hoTW.g{;u|5)ٯ^2~8ϝXk`ŧco=uDb[G.%P箌GGCFHWc)&u/Y_\h,3IHE vl9vFΆcB}1~KyHs]DvO[f #nXm\}moށ sc@j%q9HW"EY$ɭL鲰+b*XVLsݮ LN:Yε; ɪ!sr2*ujg}r<yݾ㮐(^o_S4~EY>g8_YХiǗпݎ\5Rj4m-bI\4i&JS YHt8KǭvdWх0](jE"\5Hn}϶۴t /"I[phwկU{fKj_~H&6 I %GArp@0% K礈r&M0I+CR8oO e؂jÕ٧}" 7g!Wfdw~3`LvO $;Y бEpY*@E[p%>DWzNV=+"`oઐU]B1;4'pUc]ؾUVuU\0?5%G5O[]%l>(>U*sW}9MNQ2@R(V{s 72h,GpHlY3CmiU04]Tx.벒]͎.MMk+@94ӝmZ#K(-,xH&JRԠhL :df+~٭C$\‹F껃iʂz3s==ǀޟpٻ6$ .c!fHw^Ę"$eYy!qHjJ< b33==5]տg/';}9ٹˉZqjDˊc[NfJ ۔1W_͑hX/4e*Z2MiAG.4 < ϤyLTv&M*O-A#vLђ?&+4B< lm2tiZ߬6><ٟK%(AxJ?THmtM;`+d =!"}dh CR(SQF$#89xw3e^3'X/:xюϛ;e*7=;kt63ORKo3unIsݕ!N_nnf7r'`e)a[(V4a-4UթTFWUɩ+5%[De0)Fh{JUwƶvKm>q9L?i:P8F YK۫x!>-Oo'yke-e` 0GM,G܀̨6xΨjgϱ7K3_'LԨ*0VnURaJgWae}2v )#0IJKU~#fcJsLiM5%O9"o$Y{i78!V,*}AGEZ]qN3VaA [ E !I%i/e;Q!(8v0)&8'?2jU.rv>늀=!S;R"Q2miaXV{u=e}C}KpLTwBVFo4B9D$hf9!ctyrx؄ A† R;f e)A@-8z<+S~AzN9!NQRkLΟ/ !m 8h6 ,^s,L04F{ ڍ7 65]4_Su6yoZjPg|7/ (hS8sc>:^Ka'l&PSBD3|9ĂlJa^iw8R$#Y@SJf41VGUiqMu-6%Vdm"\lMX;W݆b8N_n¨C;~qcobiJ=N}V%¾9H_/üKǡłZq'jZF׀֭T/(JdT~*MHr \HJ@Kz#'-%Ѐl؜lAF,X'Lǔa4^HIHt> ESzm2 ;gv ]IS2=9<ϺgerP{$!Nn|y|K²ACCЕ*+39nM*+Re<MT,-X5}cKP$.1=xy -G,jSt0¬ÂD.b ,@Yt2we2(:o;cN>D佖豉VHKDfx=Qf})_4^)c`0Bb9F q \#%,]&[7n2n^A2'Bh`#7|*~]ny<嬊M/fSnP)azHSUaUlDz7>rC<\2S"KId!vy'Q;hGU K]rI|m)-s>aUIe_C5Ň ]PRI>3Q|Z ?}j;qI- G|^>- @ikN(w7_{{p]R7=3MֶǨj"*9j>Z)t%,cSNOv)>kpn67:$Wv4w.Ɏ g+9ͽJl [\Ԛ]^?)`s1ilaкez{0َNm}4e -f}ΛX=z^i}=L-w/y5yK^?&=Ę-i K:\oֶ/|z{[۞ڶu[^Ƶ&U}mS>}gYd5!`X( H& gfp:D)LB`4"@B(X &d)I8D(K#4vJ~&:x2 ;FCT 4FO 1NĬ6ܚ;d D8DHk~  ߌ` NT[< !p\K5 1n@,#m/6Kx۔ ןdчaYסE?rv(P[.> /≁F+)PPP)&id ]JfJ&9V *Z.BQRC\^bUJHZY1 ԌD5V+)h rKNO×bl n~Yg3E&[TJ-c1kI\WKD[F,y.e2 `ʮFi2}<ӝ +/\$$YNi,HkNƾIyU|h\/r G'z}2λG#FuCKAm 0 [34\l`}sS0 5yZ9W_n,Wx;Uú8նwYW`^m2m%߻˕N 9r;?6eFF2qSقI8 &yN9JMy^^[ϩ9R|y7Kٕ(~;\[g,9 `2|nXm4<2IMK1@{Eq m ;IWۡgNţN^#NJaW&MV:QE7eQx8cy"WVL4B e&2v 58eZm4}bgӚCM1{FYp I*]t&q 8#c=R iơX+cwnA.WHuN,ka]иd~2^,tA'3p!JzY' GI͑=RXw@Q2Sq\&66AKe4`1MRf`eĮ&gx jҎCQTFm3`o٘Ўi̒B4f!dbpN:8HwãJupYi}/!3䒼 ! &HEYEȩNƹxXMu7;- 0Ƨ""VFDq@g oxB)#*uR(,'1cJ!9&񘹬[i#7NJ n6\D#!,(0Au.CǭĹ߿K=bSCOjZr(.ʸh\p&IgIN!zI{rCt Xܧh# <`up}w[ u9hXZߔ7$k?Nng?S Mo D F#&9o "iR|w=oҎknx,Ezo4(%:횐lH O$%r9X%فRCBqZՙ{*:tR}eK*mQ\@g-.1]w]ӷfT|J}fx3{J`IW FAuJFǦ[v_jTjPeh \x,khn~7zGz]jmW0(3#!Xa0r|6!8!sT2"HG>ecD@4u&h4FJBNiטH'1 ` I3 u.P{8{$N2}xKQi~g|G`tSn;~͗>Jfc aTtfi"ΗQ_bpe s\{8FLN^XBL̄6`,Y,xc4dY*EbӤGaTmd~{R)Dr,`U3Dzh8)r(fM KAd2xS2D}YQΪv?_m.ܠ]gm •B SGg`-"*mdLe+h" 4?=p_IMlJr )eYid}af>% l<|TL_u:ЄjQ~4mEs/N`9H`d0 NfR0X9wpwVJZu@kq>Lcb{*>H>,(Ԗ4x E/9QA9+!:.Qam\S^N$v!0f[%6-Gmr扐doy.^.*ư8w!I=,F<4԰9\2_l{xx7nƶ͢}\LqFP>] $DOHtrrq})#r8 {]ECn[(ӿV4W+%TiތQwy ѯmķ٧X.is<͓wݳ^ M\-v/޹Ndȟ6;Gv ;A\Vd1Mg-F,OA/.7!Q<&w^.mG]mic#e`ElW_yT"fݾOp|7|&;= woOߞկ/O)<}_/N_fXPK!'CڰZCx=Pq?jq_}:%iqέ fo"6GYl5Ʃ,iGHl2]x1 hc[Yl!CFv-c~_Vɯ$#]ewA$ɿr? q Y"s,赈6X4yL> R,(cP\p)Dy>way=qwk"( tr! Ü:2 ]*kQeG"^మSe]Wt I}=duɧ|t%aZ5-ORK-5ə 6lP 4)dWweC.xS4lK௚-.|)<3)0U=uYy;@з n64J] 9a"b1)6$D .w|ACWtwVFr0> Mݙ*̝B.}#S.RNY: [.q >e"G'#ca1@L?Z!J<ʤ8{tC9v(;)3g{tz M[\|I>Qm!RVgGg.١D<Ixtl6[KUG0BK6oJ}ɆgF聺wik{CK[ Bcea<3_9J43Qˈ9DNnFkEZ TY,8X9Ecil9X NBr(t`\f2s&Ξw9'xT4sۛmHuO* qYT5*W{ɖ9H&^ f$1Dޟ.]S)Ʋb\w6X29Md9:2&Cg',cqcT‘Qjʫ&|-x>-u'k__w{_St wgHegY>2I0O;YK_Rcy"z3{ڟϟuY~#/*ӣӡ?O`a\fdL[On/gz?~*4[R.Ғ좕L3AQF>Z;k'+l]:KN:7kP@_&E\t: ]_z՗}mLK=cfS$m9܈hwq_'CKڣOg~ގ6/jsvtޞ,W`WW]QQ&ѵ|鸬il5ߊJq),f~U>*Ң;\)++Rjl8fUWEZ}"\}pe9M%ZPjHfzed:2%F,%NR5eۯ/gHOj8d3#pDN_|p#V+b֪h܆;tPHAH@fbjBM@ۤ 4F'w Чer8J[eT\k-~oeݍRkBT>۫? V'i#ݔqRچq:;l\?!PFdxC>7wPD>/7=_!*+dઈUPgR) J0I~ZçelMXڋSlX-+|#5jE[ Sۋtn= ͻmw}|]?aty_ s1|{&spfbC7ZS}F9r-2z)K,Fo1*@BiGgA=\)~?=ٻS^M"'-3}r]ѳ0{]/!RյQ7-eEnЛ:*wuo߶6jy'j%h'ܛtԃp2thoC*Z|R*v:xKǾ%' >Sc{&￵jϲ2~>Qy##?r{}W8k^lvQ<Rm,Mձj_^M;h4EK,I/sV:)DJ.%R"SC*:%Q (=QZ6!F/& E̴H2 D"nZԌ>ު~R`Qlb~<Ӭb1&YL NPQ1N'h ZeڎMtۺZ8(-K@LA7/S$BDFsSȹ[PgJϢf|gϢeqf=.:;}x7oS%U^y}۽*̓>7\NٜV}n8ǴQ80|w$]vHE;N&R$%,ctB`ٗ"JZ\T3Zt%o&4.!XqY4$֚9wkvX.lBXA>.\STm~^$-Na[i>Ҳ͓ s)Sr&HȮ'WEE #jo>ATc+TUg/d*FPPզ.Yk1uZ%Qac݌5vNy0Zw쫵]c= [}6lry_bQ2K'S:X!AȌdo c]H|_Cfd(5{I.I$ 2pdBdة&ZmhFwkqƧojDlqЈF㥊Nj- "1*0&ڹ$!XG2ih%}wP9&(V:E{Ҡ:ԩ&5ֈ١5kWY/κpR}c88>FU, xf#{#KAZF> àB/>l{C%>Z> [o^9mG [idw' ُ(QAUcoJ\gwBg߿;-wtt!Pdq5;>@?&xpM5"%7FguP]'VK J%\8.%mt ĺ0J[N+y[der1 Waܖ*t+rK|HMqI߾[a 7O߾%OZlmÚ__)+ގ^/VM3OWkkWYs-fBkP@C s]C+ԇoB6|mWVUھ3؏.{!"U(0,R Jltq.d4x"YrF'@vm.YtNkI)`%Z$GH3u`֋!1uIo#l_ƅ)ߒ;X ma+w(O,іAV3Dha.AqM喔j7i1xѦ Dic03Ba/ʠFLJKO0Hn;]$xcsFE AX X0>P3?k=O XM, ɌLuݬ BJ/ cPIUc9kFΎr}b ?Bv25;UHVIs UIIy]X3IQbCh){RQ?1mMiADTC. }.6x~9$B)bsE@oRjt?i-]@$m<0CPa`Hl$2:eL@([ F5S9(&cFw֌'LsیdHs[^nf O(J2BL hik}tX*RO\+%#L0mg?~E̞f,·&ŋCY⊱ڰPݟ&4@(J5kt`BBPfvcU}+%Agհp=y/KU#Ȝ^.n=D8Y%$ &(9'Gj_Q Պk..;0DIMÐҌM@1pPH@ൈȨknE<(d@iCr[TjcR8'YiW={F L`cCcʲfXuM)Ī؅rᜍGJB,)JkEQ%ZK_i!L>bZ!M96<.j+f/: E9UvqršA$0 ,hJ;pq+=Cմ̼xV(d1&Y3ه֋Y:dH9k- -缆 9'a|?4TK[xpba,BI 26:FŘE)f`qO̯{PegZ}_'ʶ_~9h Ɨ1ĐaKm0f:4j{{SM::Pb (O"^'uV|4}wjdX*O M ]H{~]|=jXc./x~XMbc uTzmn}KQgu(dl}:=K3_eT#H}X2O}qdqz^3;闟^z/}շ_٫?|VO<sJA<{{=6~}h'Z 㭆VwC˧^N#_Z>q>8mk/@J|xjI~գ.\Xfc aGy&3x6J[T*U-6*C-c>P)/%#]F'IZ~~"FDv>yI=A$! 9 l ~pXq9UrP.W\ESVN%m!WVѢ)&و a]ɺA6$u՝Gsk jH4;hä̖YgN4J:;bTTJ&], I)c$YI>, ,q!aTI[marpH%.Oƀ22 mQ"00"?&};v`gC 5cC }& n O35l[/s2/ st> vbLWKujѐS6$7D w.wQe`P>+:w'kk;ҩMwa[\d02B*oy&dSDPt/{WƑ?>y/nFЧĘ"$%E19xCԐ WuvlCzķA2hL^ w[,Rlˠ5rvNmCh." |互~멫 d^-jy4¢Gˮ(eM8vБj^"=2 e)DD%اw*XW߾*"SCуK?I "S6O=ZKOR/dSeQ,%3,Y!șķ 0V^W` .TxD ޺Z5]cmKeٰCy ee PqwgKo CVv~^y_Խ_POT}7+b3`R2sxAHm:E`{]ͪ`5fxxK0_O=֕ak:uoy [GTRfrIhl8"\ m8J$Nd$M 읐&RiD~b\ Rk*:|mK*Ljd<)V֧s^jxhA #(H8HFΎmck\揇SsFѸwWݮH֢z-vz]dޱHl[J齵VEM`ʨa;aK+iRE, nD:><(r$z8y-XhmaTetۨQ1,%! j`SE9wZp8M{/v}]0m+:C3'DYp6V3O44\ΙV$7<}jLAua4Iu„I\h5o M:ه[)[^+uAl;Ze?Vй[S|ȒmoFr/t:ڣe_>_ͨ]Eu쵲2TdbDAG^a4UTsc%bԀ 9NxA'm/_AWvPi*Ӭ`v@TgRaB&yW5\o)Y O7$܂J$ɕ.:"# Gg Ef'LN舜B½r gQ):JG,UHvZ@ JHp7 #$ pԖi=~qGcg^v癠TAI9[{5F] `hn)1H4Rג)E 0;=)z9Tg*'&=%d0gWp >ىukWCP vCq'p^N( FFQAgT(ھB I~,Tџ6r;)8'8МIKl-4.{ܡ.{FP!@$Ux! wHrCHKY\G&:ѝnf7>%-A<%s=z=.O.AcAgPԾnx7ku匛frɴئ JƋYXWg^*#sވ@F-ǁG+DCqLJg vVZ=6shycӋDw3/Lg}%vIަCljnݯZ;_vSw[:ytWCP˿xsE%qV8A˜ysT-V1!wuNB3BOE3ݓ|Is*8,H"Vj9j;A+ bTd2Y!e!xpn)TeL#XwjnD j[#gGM(ا\OG R`6>r@'9fR6T;V力pF7 Gٯ[@<G=:ddRԹ&cW Tn| ?K©8f5p)'XbB)>2Ls>F^IT&/`'"J5>\>OAQ\YC1ه]P `>k3- 5+?Pz ҕ>N$dj* "]Te^&)nVMm_mK ٰm po 9]yxmlWwozDwQqގZ ovlJsd7R5IMI"/kW?*ΎV{%H%FӞ!(Br !K>.1WxL N9UWWa" $18ÞA$, BP!9FZDWt{8B锂F~z{vO`nynx+&=b).%%9~pB1 &i 2##Z;WS20j.1O@q3j6p%a&(y8S[p+nmsx!Mjx[[;fJ|dƃ' D`/ Z)^8e#er`O>XG,cF1vp<id=8ZcFG$ʠ#_p9EAHE  J^0q=3x񻞐D#Z9j`"G "c*Ñ]ڮ?ɂ'ͺ - 6> vk݉l /=nTn_8Oi ""KѬ KuCL$6ə ]Pr2كC%;F8P:@j9f$ ʨZIAIT=-⇟s|=y_w'e k⤾n7▱(KFJEp)3Sv}?܎֟AY_xENI%y4FȌh^dNQ"%ƙf]妗R5{%}Y}\Iy}"u 2%1 U/ h86gx5b i왜 Ch5}ꥧϦweJ %τ~ ƌ>Li\}<=6_~^?_g `Ǖ^' P.3U7[|+3Y~6vJǧ)g4IMF.4S:1S:ɒ:3i",D..+ "T~NWaڂ("Ibd%Cm هߵ]Y [l{;HºpN$.S ={wB8G}s~5]f0i?n߾:n-طǾ?ˡ5g - QӊތJb7M/M,m_x7+^XKDxʗPiU: ߎ XCUnaF+w•uv]u s"E@*B6)W?yzN`L8Ippput`p+'{TR0KlK-bDT\RK.^%܅n whنO_~("ٿJon* .%ym~r-m[_ 9?IlFUt^a4:_Ue߼ߋ=83Cٰ3]A苿 Sw}wG} [b2n{cp!e[`t+bŖӕt^YK0* ޘZ5<0IU$x0@;ŋ`GWRpR.*>Maǘ^;6֝Y~_Wjug޵q$0یՏꗀ.C/B?%#NUI4&cӬ櫪@IH!ĆH`kJ1!ȍ0Pѓh\]fޒN[F5_!NDWxNW#(.gS gPpsgYR+/y#PD'j8vJWOE;R5umlm󺒝cG> >{\yK .ljsqpUZ.kdQig' 'kOWsڣJ Y{vY{^aMsRW ٨, 碮]]e) +TWZ5Hi$rݷ:\n MHr>BD CH]z?ޤ">MϱG4`QyI%'o WTQ.#ƭZ^is6j:ع,-Uf)kTLIS7挿r{S; nyv y7 oA3, 7G!wa1QkK fβ, C1q1Lڟ{<0:{PPaIZб7~Qe]ҽ~ʷ wI!M3unfO=uf9&(AsEbI(Px0I6Ft$v,R-,-"T'ptTgYxf?K[4{j':fO)M,E#:ЩSJH]eJ8uU*KSWG]1&(='t~dMR\}.*KZv+礮P` lUWsQWYZ)ڮw5+-?]igeyEL$Dor]oWA%9E1j*e\sGsC\"<\:b*_@ x[ؔ&{H#Yo V/5Vq7fy3/ n aǟRro#'_z@xIub~v[dPH‰@ByTVp)Khw >ZWNU YA՗<(/X0N0yHZe4IypN$z TrlM,O (DqŜi80DP RE>$iat_Ǧjgrh[;*$}# 8 h⏫ˋC7Dft/-)/K2(IhJvMGș\V3p NMdрiOHF #fB)jc$6E"0  ^iL= NiNʩC#K"aToc@p+[ M.@D1,ּK\cI4 Z.wxI>t;xͶv+&y:x6=f j6-XŜH\PSqJF;Mpg "4i -k빱R@<$2A£ZĜCSSj:[Vc< j6 <0T퐶Jlͧߒk,U- zxg!A>N-A fp>2Y\υ խsf)R 9v ms l1 B 3[HJOшll\6<f1i2(d^'4lRL:FR"FfƝ-: buW5 Ǩ'&[wI &-_0yU>es]S7|5pq4Ie "A2P"1aB0&C326M&z`N{(l;1(r:XtJHbUc OaV[,Rg-ךROmh]TwItOZ"fD4"NZLV ܯ,զ*'pM\K*cbo0eO7q؛a|:!PzƇ\9!nspiXSQ%˾juvlv$ ]]#O,CFOPjz!H`C0%KVR>fĽQ;Z/LBd".i"׋f1b Wi#E"ECE  RY^ RLiqٽ0@紛 5tqOCw9<󎶔6-s;nߞO6r\s|t d&]F;Pv &(js33S ORv4vP@zb'<P!D$F!Pܲ #eԨ%3(o 8P0^fI 5 &N &dHfEyDym"aژ8;-Nx2Ec |䍤d$y Er)tt%rgR; \S@7n_C7:gbNΕX)ҭ RB~ٮ/djAN%dqoG T3 BFPB#w,U}*|)SUAIټz2UNe6sy9ɬF-$6Kq\ӳbovܬ:F;AZR0IC21)Zx\s#,mr1^gաuYrn :1Q5s0{ v& 0FHOe*5kdAQE :N4X0v&˩]cw#-/.HB~)3A dIȔT#TV[*"XR$v]93nQHF%SrB-#TɆ򍉳cyh][ \! hH E / B+~l¦Ȅ˕4ӟ$?ܕ+ny4㡓]2tr1spLn\p*b82?Gn|=!^ .b"9rj D*$Pī<?4Y()6N#qiЫKX!'o̊t`.dL|R7PNv} > P5 ('<%JɢXU(_?CIPBBBRGBYe_m~g T^2NC(w7y8 .[9)ĻQ}ެYYe8M=5t-SvҠ?ͲClY`ˢZw>yΫ|d;rzF[[\ݞJ{5]<*2t':<-`[/bf'a|qsP?uG'xmy^&҄?}.~] Jv~Wwm g 8gp ($(ML@*YtQ׀01.uH͐CM0b;ϑhPg`~E *`D x1iK)R`@s ?Ƭ73V8SFYNZ~HjbPbx5y"\6C  (l>Ep]P#BC\6kޕ)"y'xt[;zs tzfC]`K" xr*@ԁvSHx,c ;ӌMf7S=<_Ѩb&Dq6AFJ,~)\4OSѸy_?ȕ9F)g qVE(7R*XPrH4s4QoҎrٳ&CCO (jڲ뻧\P|v!Ԅj Pξr5SyJ)5:$RSZb6њF|urBM@bC(ӁzgnvgyK ڜM=SOK= @3E&z#$ߔ+<-G?˵EƗ(7 "Xy]B <S\7VBO,el\Wg:h!}0CNc\/BAYL˹\`C}WVlQR~p!~I&e=1hE(~ӻ|{Io'IUv}c3\nKfVҬx/~s2^2ƗǼq.pg;C`Ov>~OU?L/q}^~;\_ܽ OkT:Bz]]oWUި]]ֱzCDQ6(SvZTrq~XxU~.D?r&vzh43JCJ-rtytK;.^;ͺS ,DCse"}fOK(IB;)3il>zMi2H*re" T$lWe 1z=+WxN@)3au@!12"V^ FeOpzvJWOEC=R՚[\dGGqHswv8<~Y.ąUmT_fg* ?{WǍ_ؖ΢s6Ag_6HVQf4H߯"HeVӬfSb]NC:9C:u&ׅ!ʘq{MۣiG߶\1[쓽s;!4*L1ZB- $dE'Fo9lr؆ GWN/k&n*;++ʉ&y޾Y_ȺЋåy{-}zwoj ll>tmۖ{n0<hlGHHRgHkRx\'x+0| P6=/g0qڧѭX`-ɴE'URT%/Ug5]p>v $u-,%:B 9A RUKM B%Y:_kkǶDߖF&;Tׂ=Eeud9H6Elv̙:b3iɑ>%U,uSA%#|P"T2v27d WL'.YL ȍ9w *,zcXJ !S|j͛)k&m>2Mݛ聊w&BHDGC)bdٗ"R jF Dx;,Z*V=!XBR =1r:Ji [kfܭafg ]c]F]i /*2ޑ:g[e>.wO.NsrfʐѨĮ']E'PToNLbdB۝tETBjS;IΠW}aN@(2LEd5v3ri>+.`v+8TkZۏZ{DlB(g B@s&:!  , ||_iCfd( {Yg]$S1pdbbQdvX6#n};U1ƧojDhaԈFNN^#"VQZ˜@2kc>XK !;$i-Cq^jFg uI rfOZN"M n5b3rֈo_zq!3Xg3.9T/z1zqԋq|J |yv !Ǩ)(T.#J0ddCšfq>Ԣ>\; 6hg E5- n~|G~j.)asqrdLr7 lLlJˮPg]4!v#*%jw||0v0va8>*)$KnJ)kM2])dbdAV %h3`-.QP#$I21viLJ[!2|j/Xmti [O^/17vyëo߂j"asJ2zusČxCsBU[/@ERĔ TĺV@) P =l&ۀVUھ3uOp]BE@H+@T|Q#E>NTdA/zmO>;I]θA B"֧Z/^oFC';ꕺ#l_N S#wz;)w߻.4>}0,U'hZU^)6/&3 P(%I ;}I'j Eh#QrI_ԗ'=)Qc]$$Dᘏ` Bf~{f5PY%jj5!I5PŨBJP^@"Ϡ:1r֌=嬗luFq?,ej\R3X ŋUʄ1eеeP!u IQfCGh){ZQqUFE\RYX\ 1f)/с`s#jtF?i+.{ IlUԅadHl$(s-СۆKVFjrTMǠ4ܮF[ }|7Ss'HgTtLdBhHD Zyh[FDuo" i,|l}8ЀZ_ei@\!l5VȠ6[JΎN<`lPwqȡѸ_cbY k|I_dNKw@fldGJ6aE)ۗdu 0ZRX'ȓn"O dffYָϲq4GAQ{sFx(.tώh 94 A)R1o}h\|rwE4;K~_ 'q2}on&d=K<7kg. uP(uMG[l+^ג{K(sN=^?0ϫ3/\^xq}HL מ?Ngezzz5:\ 쇋ًq)͞!nis$col0j0a2+DټQu4Q~zP?ȯORǤbŀ=)HNcAc)"l!)Ue>uad3t6\" 3j/.+b."][lv ' NuA6/$՝>M罇 ;71uQRw`a \g ,:(G@;dN3V.QѝO:XL0&I)O.FN$fXVn²U\}<{vO'D-d5/EjVYlF&|ٶY-"# #k]r3|(f 2j8{!恹K٬=}@g! n!w֞9eR:F15`k6$7D .w|ACWtO־FQ`HS6݅mURcʤr}Q؈V?DMmޑL1j2 EQ[f ԘMf&R1iF6SHCj[d }%LݺmFw÷gY/[YeGtHHyW85[(' l>GYn#B{mbA)1ud'`jC4£ޥ ͇P7.mQUNbv>"NxEd%:Pr vrγ̷_+F猆lXMbh k='vET4"he1#f^7kmnH/71y?Xʵw];.;Ηׅ̘">dn >DT3詓G^*ƓubYcNEg2Ř^P~a]T.0 MjN-s`&x˩z'eR= ֍Ϻh}4,_}vZ2]!`E u!u i骁te:DWQ ]!\\+DX QjURMR`k\)Bbe(YLSxqD$YSxMFh ~R#} P=JB0"˷wo7(Q?ȓeMa-;`gC邂3YFduu>_n@[l1(&AEXi7D_) rB)Yrσ @KIwSLYuA.s>4 C蹄طZ ?22L;,*.OR %?iIx|^r&ڏnέd[A_"WYv&#Y`0x)90Vh>#i1! @p-JC;W{BL5.(fh+`ݏ4y7(FG!Y"Z5f}iYxby>hD( -!,K'6 JKKbk~@l>}oOk- RHs[,ugi⬎rT)Ձi2)o6N! obb( ׎S=_?gg*-Q4ʲeG,ɲ6ąOȬ8v韓eZ3~:R*9_}v>vpWFMRwM48]ASM_ѩw{$BHOu8Z2vE^Z6U2Ʈ eGZ26͍/d=!#3yl9 w@!\́@~QRj(T-sV;^ÜB삓e(rVVh7ӂ#<:3];.iҏԘ0_4LFw7q-n%d B8Rjj)kZ)הo,$t&KL$p3YiDiHDrzK.ќ]5F*eIϫ==JKEW]閮v=R1  ]\M+th;]!J[j ]1m6SW [wBHWƘCt5u!%'ҕ0L]!`ɜ+ u+Di`t%8DWX0wpƻB ۽-]= ])ѕ2Pw.eEWVUҴt@ҖRF+|,j ]!ZkNWֻj"]+Sv'D>W -^u+DTKW +K-݄]TzT8׺\XK_v3%9YTyv DS?5ߴ`+-=iIm<^˜ePDFs|~V(B9aP Oo0[{R~6^מxd%hʺѕ@W]E1D;DWXPg J ]!ZNWRHWpmCt 3Bԝ2ҕ x! BRם-]5$9yW ОlRFCW2ICt; •xWV˺[⹖4Z) wu-w(y;D2ȍ8DWXw62 \*]+DkOWR`R%JSAӔr-3l~ `~S ̊(A04h#ii.+IUj`ܙ+*5LJZ닲̖GL~Րir M%J[ɺxT%)^Rv>8c1Q`eO}\mN(ZNtJ+jvVhl*z\g!BR9CWWiW*Q{BVt@Z]`ƨ3tptrVCY--]= ]qN>.jp)qsOWUJpfKt-wރ\+D{=PrUJr)8u]!\f\+D+D Q*UJqrR S퀒V[j]iWRW Ѫ{WRHWFECtt;++RHWOZ>wGIha]F2h_v3JIdjsArwIdTPHqHW~m+YBLo@ġvK$'OrY 0;5]UiZaNJEWlOWyٖv=]Qɩ圸BWV0SwHWLCt9%µJt@Jm+q.ծet(n骁t%T9DWXYg  ]!Z[`PJ]5H +w+kBSUJQ΍p]!\A]+D+k?w(UKW+ijj= н>~O/_v'S@:Ђ}MTK'+oo9I3?E_I^f 1oC- {5|ZȬu0޾-72QGA7>+l\ȟFa[ds>(OI%8vbA Ƌ^\9ȸ ^=_+T vrCnJL:^ 4Te͓=V?$v>ƞw)-z!S~Ŀn>Bym(co4%FX:`}JY/i]S4Ѣti5 v;ޣ=:ԧTZY)]66sӏkԚ)cP .)EM>E0$M>t.%IdOS&lt6,|R~0 x=xv|I&^3#M /XahƖJ%8 B\LldX^D| P J3׏#§xH|̑ e -8`?04QOt揃D#*D- H8#IsQh92 G½hX.d|D_P#/KDŸMƃB.KC)e#'TehIF:1io)?B@bˢTVeՙfI!h|I[-pȽ̔n|\!2ȥdyRsC_0bmb=EVX+/R s0? &r 60RZD70"O&G+V2)^15U¨ĸCW`$OgvxoE!e;JF9[B&,qJ !Cw rn.Dj!I$? EC؄"_W/;:; )^,B`X?{Fn /UzتVoey9N*Q&E.v*R$ r/C|FFw#p)aSth{La5fǏ3վy'Pyy$@P_=&H.O{ jw:xyB\QSCe$N#57ߣt6b=rH!f?pJ 0b0J椽Nu4V>Qev1Šp:ȵnx+4f׸6TH7(A0܎XFY);ǑXNKw3ndCtY{ZF sVfDt*Wvq(f1ѸD5p<'!>6$Y rZʴAjF&ӻ1g3ȅb. %"I8A(O q6D&ea/@8+#eIJl RT9Bb)8~2ˆ 'l˲Mgg˲xtIȶ:@ e0wkݲy9{)Dm\#Hv=[3V+Ūj. dԖY$p%3n:9gZ |oiis@)խ{;@ý.ͷ1]egjRGu2& KF)hKMjgՁ>5py 6މf#4MF z/Gl] }eTެr+n/|;hSC̟ %䄋Vu=lboRLUQ6q2}ga'tIeR01 rz͎ûkG,K^7;,% z qꈫy,IJj ye9i/V0lMO;ҵ,'e1:b㬖%( nsVBmb ! m),q>%4H礼8" G`mKݖ"a=g 8Ih IJ>{-eMs@ʾC9e])kq0!nX0XStƲ#q !DE56$YH{.7bYLdy.Y-kKNsVjY.Ke0>݀a P< jYY"O4 qs2CpV72sRRj*iP˚omjY@,BaC݀Ai.nYa@{$d`zcjl憐Ҿ~ZS?+%*nKEG4}ղЈ ɂ)w *G0x5C@}/0錓q߁+:$ d9-(7=jYW"िmۜղ"NB~_/VFjuX~?_? _ӏFj:/bg$,L,\dE*Ɯ+8 ѲLfJ PD(EECy{G}3j&`bu9<, ;\ج}G{WjZhG+q˃8cpaˡ LZ$e{L)w 5=W֒{`˞^6}Nw[❄G^_E'vb% a]@y! .DB☱Ih4-R &'i&x ARJ1جjpBu )T 9ﳡ6? |>IgǴf;~K'S_"NYTYbhٝ^=,?=ȖȾrF1gjBDR>~X}b'y{5O̫Mn _oǧCtgR~Ւ>wad xgT3狕t9kVJ$ (o&g(:)1B "Ly&RW") U|:Wn0h݋PKdHއEa vY?`/(1 <8( icen*p(R0*UL߯ӧPu+v K,6) vܗq+5b/y8Ku׳ eo$HGP}łhZb@SQZ AI,$'HDX`){ Os28鳏:vA&,f, (ZVW_lIޱ ѱsQ%E,] Łmi gm {ղbU; 4wҳp!ks8ˋq.&3Xn}s3wY t2&WhujHIt(nR1ʆ0($)D-* ͦ,`J6)&{k8ʦ` -. J [x%yDD@}\093NL2R$AP*2lZ1D&Pn3WVٿ|l~E#iߑgYD$`L(z A]!M<@lz) Jfވ,itw^TҚŒs5+$/(Q0 ɥ>C l<;r IRwSU5;}Wē$Q2Iy%"15lZ$H"*%qkFF|hq;|) =1Ӵe,s/Te|Vъlz=<⒒ݑHYӥX<Vi7kkuO`$҂(W, $\ь  Ob\ܞQjBk,^,8QðCL fJ;'7|rHe$G=)kӃ NCfWO2 f{vRmkC1I}Bgp21F)Y:dps#@qA[ F(!?aqX̴2eDq5nQk573 EK_!)jim6ZDoP베4DaqrXze4Jy-& \6u=§| rWThhzLsFHrҵYU!^HJ IT"$NBM*fVpy+B!@utz +"*䙔RD7oYC ߫Ƶ !PxB w6_`ԣGV|:}1Ar EqH z 3jj'f5<17SHbQ3Y,/B`!s;ݗ@ħ"NY ,{ Ԁnqp %z&d@\NPdՕmѣY~1/fs:6_8') IEs&AW߉Hd^]>! 6[>9>V`Mm{! h6&]c4<FGƲ퓄 mn5oɕ$*Huɚr yk@':2!7FOKy6"K8OH4KCL,0A^si*ZȌA$"ׯ^s-յaq㟸#E["!A],fSQs3-jziMZ*DH# }R즳t UQxa@i[|aH7CuRHIjVj **!MSZd7rTs鵓76 B1k|62pD&3)g\6>,Qp۩ gTgY=0)y~Iz[_mO*v٨Q~Y,4_$_u4xʗ>/YtnOLۋ@Y!HTYDr&/qT ,V]َ;^D%-԰*DVV[N}UkIB֞tv }jz{IoHYp.1ZH @2#(Q+X%in_ ˰?봇}]kl#ڰ|e;Yʊ$Fij.QB@0Fq.<&it7v%+"ٽ"lݥV;L_$}%fܬDТBGqyQ֑<[~~uF04A" O3-!Z݇&IAV3OY=[.J;{kTQLݚ\;OU$\\^H+q%4D>:Fu9/"}˩ym GA2F=?: ;|t;Q|rIp,_%Ԕń#3V/F/x7)jOMכj}^{\y/{%3ԯ!\P?'PFISV{us۾2v+c)A]"Eq6SReudgY?4T DDŽ˒6U"\*rC+ֺ6^~ e7;+>@aQ' W d6tQ*g5-UkUd qa;Kr ob0M+4jlM`** I[vG/]bhXXqPl쏃Y[)z"~<>0”8lG&BejGÿ́PG0nr",+~mK>G!2ƴ(ʅٰ'n 9U`FHFhJ?γVօNU\ M\6 d_ nﶳB :enmS7mx l,Jk;-HC"C3R( C80f( ]QrW{aԡγ+(գqғ:*wf>vxH?Г!Mq464PN:!pnϳ"AE+(KTz&M8.wﻳЊH,6r$nҁ/"V3ݦt D:vXGA*)Z;s6É\aO~ 51jw'k߹ \(cfEf;7RNjJ9;`~ki|ܯ* $8֡Px.^ldQ"(S踏k8g!g  *WNx&яhHfA4=5#S]X|?RC t.wy5[&ԿXe^V4PX5ܤkyv4)5E:1PX@0 XFi5: %sNThTo <`ZwUitxK5ָ]3M_(!P)ĨE*)IY~e9W@JG`-$ўTkFkbyF7zi_um&-`rSS UQI$f@a dah@C~qQ=qP euvDĥu@E<&VqsaR=`ACWD `-D !!^c`Сћ ;VYxh{&񯶇!Or|%ox~dO%qreDaWuNι { \he#3F lYra IyF>R3R )Q>QW\*K׃ɑ3V1$1ꭘʷHE.nq@-$iOl>;]J,l/勳LQ"t.D'i 5Jqn?[59mϒUɃ\\DM`VJbiTޜnWh0?`FqeXi1xJMOj*O#T*tmgl#:Le2rkӤ' c>Ӛ+k)t? s2q !A&C!83x돋F>{ܝ].+u&¹U͏=UNQ0񑕏 ~%EJvC &ga4sFz¦!|a9e&,˝xL38X* =0feneP\|9emb(ܣ,8Y R8'rcㆅ^pdGhsilBݹɛvLffoϖ|aB250,ݱ`K02b!tXxZظ? 0~8|Ad5UNeGeKYRiL$Z1!ue{i Ҫ ]qx:rʼn&Ăҡo%CE(eġXeacpZS&xe ZR/uCE'n*jRp|dq%whס @AtoEbS췐$ #|bngBm~ 578D4F ޘcʺ޽cZ/J,VU]`w!%(#=y|ASׯ?54>3%D:  Å0U|[ж_  h1x\v.ShPDhPXxXωdcq0gD#lx+j49[ERw?\Ifء,OAkO(@-#1Z`EL{̓cBejDFuux~t*ұ@gl#{i2ϑ<ߤ\6.{V/~h{N[byKs.hlNBLa쐟f=,8\5T۬cmTX#6ۡ %k1=Gy@-}G6}}JzHJ K] #8O(WbŤGͱ8lx{7G u.G#4α}%)*FۿB;nj^6hc3 KָvJAHCP&֘Te+Y?hƺMU:hz38kUiQSqRidaXy3{`^6/bIbKYc2dx3Pn PMJPF9_e+Ym{ޒ]#7=Q-dn/,a#?2L0'q~}FZnN%/Lpmu7ދ +Yu tIƓJuv9nu](h۪ytk((i{rSW:+rh*iөNX]TNn\]`Xl/l-ų u=# m"j#jx$@.W,k"B XajB{@un`E:%Cx]k]F8fi ZmN;}VJh1v& hY1ʻy]h} 4w/[7KoHYT(Pi/oZ6JE][И]ij*whU"HᐛmF׀Jir0"AR%ijppц^5S] } sP8%2JfG9)4DR3e1ՙFYe?"DfK { ; ܭMUC?tKB\狊Dn0>Z SBaIHܔ$N4E@}S8`QoޣBnAq5kQELDƽ~*L ^D*B}?(0Fi8<.(­* Jabg!zwm,t^Ot 6{3>0hq"_bljsO-73n|79)m6d)T*g몬"3΋.ͼsfS'PxAw[ٻ*Sc?zX0 '*ǧd=n#FVhOI>feb,Sof|S`88 _z/,-w}̷!_$,MIvqWcxo0ߒz_M0A|Oh-4NWLX?+g+*SNm58ƥ୺F3p4E\(}\?c~KP >*;dW>*YN)0yungT)amjOĎU_g$Yi}1#~ 'g07K;g%ȢA/yNYc|><>7D/q~﹛=wS)قINc5 T9XanBD&&PaFiLHo˄/_sa:Fpz7?.- Ecn>{2LKilf/$E_٠7\>[+Gt mK. ܢ+4Y0E|g`Ms&._,XߜfE|g`[}q,S.фo?_f/_޵Vt"VOO/2F~oci}[u{5dv߭fXÝaϐGNk~HjaϧjfxqR5>B7z!Wi:Bf.BM̪RxD 6֯:;=8q/NiUWrmIc J;!$kVi(?\1IcћTǭR5n\ Mgh޵57cڧ02UyC^N7E[n-ɗTJMI"%Sx9c:y`\~7n*DMS3M,':{jp-koϿ,Ͳ_#OP[R+D)TbCTdQ FV&\km(+,u$nG7Qx<qi{hy^hZVg?-sK<!(A L-LeiP7 r"vVKD|P1qxSmEh2 wx zm2yJkfF% Y3*`lhkyx̍e$$S4I61W>yCos)7V눱6i9rLQ)/FLu:V`ӄSҖx$ө6 j%!D*$5e+@w$47 ۱`Z8P>S9} D#b bU'S ҃KmJ2c506-!!q(%_3q '䧤(}vSf <%G.WE_ˠƙ1C\+X-e@[#iËވ?;cp*>h=y(֟#Sqݡ_"`JspUkP8jZE'%ΌVgFVsOrkwlkn6 x,)}U•h`Oш[;YdHK vVaF/ĎA'Ԉ2"j0aDdz+Qs,ʌֈkkQAN|עmpdphrXAd pA)M\G){\_RQ0gl0pSI1`lDa3+3$Zk0vvZ k"Q<5WRcSeT}q@EK*DSxO8䌎G9oè2k+6ۂ.GyʮOFw/PvޓՐ3:HYc5W4]VԷ'\61»dVvkQ*-/'8>KgQ6yyΧz&7}k0I`ҥ"zL.M.}0%2أ{x_2Xty-m՜-.,*Tu| yF]_>]aLb$h$8s~y;V'7=y2Щ2&z`Z!JZKcjeH Q.Yfr]aU>x6,ėlj? aL׶I9WݦNm,J/zey@"]&prO.wI *zlJcM~*K'@_w锟9dc_8x(( p^8%IM˭oQ,6gEoѝcog~ |FE J=QL%X N3ow3ɌњK|/yE?ƣr"Y׻x o_q1?CÚx`}EoĜ$j4E2>zQm-|]x}7"D] YF0*) G̓ʬ2Nifo:r+ F)ޖ_yHJT.Ps$(y!Qx}^ϔ}|n+m%S\ӻs+< =H~iјxBgJMNsLnfT;Cetk&Tsr1dAU(O^T+M ZD:WŠU'e >K惇*S+6FӀ' ˑ"ɉLQVvp'\<2<*UN>8Djb3]9{$3ps2qkJfW܈Q9Q<IlsöoEB\xsDç6n޷z`c"is(UKӻ2F۫]/Ub|hXEͨX:h{vQq1 '3P1u d8sc D^i?PV…FC8Y;5K0d$j5ߊ-+-zm\ӜLj_& 'uyi-\7%KM>)aw:YE,E79q"2dc(r!|3rGcl2K nB"\tT8yD3292 :=ȭk sߦ%b * ׈#$m]<*-fZ\z20"!sD+'lI FM𿃅8zf˽s"q<;9ύC #gOS1X+XwZFR|ٍSQȵGֲ,rm񥝶%42 GRdqL!?jMHTpdGQr1&V F?+&Xh~|>J7CnVe*7Va]o3gNX9=}{-v7EsAz<Mo)t!UEn|~FDreפwj]qݵՅ5O %Q5Q" njҎSmAS " k0.~aMRrD;= TEAFc;l8)iWHQ,@w `CrǟE3ۆ}fvݝ)J*jiEu@§Y搥!츔FY `l5z#ٔ6A m@^K92ȖQ!z[%_ L\4(4kP'p2&H}F?ƋX] FvuX7p]6[mY@pPAZrrMUM.|Ʋ露"C<GVX~gf/tRf<^i|Q o=M]䦃G,KPt ivFH d1k0z]Mkm/9Ba /F&ygklU`Y32-ϿTiMccM;YHj]TJVטe$>߅:|]藄J1v5=bG&mݧc~^[]j/{G|kE`)fO?>؄ꥉ >?< :j'w뎒oyyP n;XOӌW~.}'ǟ(qV;vѹ;٪o`r;)jn@/ςcv!ĢL/nr/6Uu ༭㒞z5^2ˡuF TWqϧy&`|w9-j(F U uH.Ap%c~䂱F=oX!鞰NM{0&fUї8"_X1MHՍi&[B` ?W nhs,F,%U:Zh\ I5{m."1fmUۥv.+&A_5:z7(|QH#IX6VX40qzdxWnɟBv$y!/,S5^fbp%L1T#Lq426^]0%/\BeǪMPr˳Vm:5}kLMRd*hJ#q]M3x)3m` feάJ P ]GFEùe )']!:.ZYW)T$?~@jCCɽ0KrFcvaP5:MqOb'(݋fQv-!oERiT cͭhpnTVDɞsQ "Ba"`Z; fGM>cͳMmqU%s\W"?~\'0 y!t=Dkxc%7aۯzF8PvA{W  ӿ~}LƓ}96*aakǴQaxXco?vm1UHi'ӬǞ}v^zu89^~zx8o&@/?-çamݤvHܴ lӟldizCCפwy}0vDH33BEL CY[x5)?. GoͽTޣSHnלg!YV sS RVR~` >.(Mkvss' &),Fd Ս;5̝`KhhRpyB'=OkU}5]nkR+J񩇃ik>h~5ڗI0ꗠ,ǽi6v0]60}qYL/+S(TM8_̺_+dHW[5(Lm(6Uߐ:?۠Z/M_{J=^W,5ͱb궝J]1P %RhxFxBIe^4QTR/ϢЊpHZŨP÷ae)c,6m@AZy}ʩlDg6KИ(I%OƇ_C6< ]_'}L[M=ԏyL"yw̵Ņm58>o[2i d4!m"S+ĵӁu{}Vb6ML Xb#w6 3*Gw' $Uo#(O׽F/[D-tkoV׫qjVg5<0 o<`A#&C##D0~k.cY %NrōHe0HQm'$01. l%(Zr0 ۔N ^@Qb_>Xx/ 5RnLq}jzcj(w) oC&LdęC<.3c̄j3QѬt15bXpPڧ?i^X)JVF<3>$`m7 (mZ/zɜ}e{846=3iZn*:>[V=:#aPꨖEn9C\[$ n04m 7^YDȕ])qP[vuȻH `K:Pr YƈRZx}a^GH]]\]KYضxy= `C=Wm=| { F$,`(C4V ΰˋ],JAٌLw)j;lHY*5*Rn,}r t[ars&ziC>\kk V-<[m 0uc[3Nй쿠mgg)7 j}S0GHgo3-0\!Ͼ(OA GK823D[Ky:O)cQ0t廖BB'\9OxRW*ɋQ:]l-r5'0-.UFk >FIzB1USo f2Q1JalLp"鵔.p\4Gil= *+ѕB{/tkeVꯞjK2Q娘q>!Jc$d$Z .Q4)mz)R]&9Vk]VwRH򧑫ʰmeÛ17Ťk/Hyl줤J_{N5za6Ա4ڈḨ{ނAˍ"fT1ɭ7,s-2Ɩ;g{ bT͚6lYOEw XeF ֑Di\ј'Zy QaGJ-$ 3-Sc<!U~s1`AWļ.($hΝN7R<רh8y^Iz/팰SBHk&5xPT{ڤ;nIiߊmi3ԴۓO,s;<|CR Rc77- Ң.y.c8%Eb+h,aF(;K^SnH.ٺ8;[IœVD#"!gT3=(T kGўTD)]ɻ/ꎱb?ӫՆ)}SPyk}JjQ9w0"U0$*hvdIJqKf^LHpBZ աs}W#bZA3S8&ɡ櫧)zj _DD >|csS4iE鶞 =Q[~d#FJ-$IlT>?)Fd4"-{=H`)mڦ j6]zXO;xMQgElŨ,FLn3sAcS,#(⫕eoMClAIݞjt%53XGmw <+ւ.5'XgLeד CS/G7ݸ5ڥ~z.7LT7!f"6Xt} m1ڵT@tMYal{"z}M I>;AT1}LvLż )|-Iע8[U/ѐ|}St}Rc^fTk8lR&L~5dk~I臰ح0YԏyEg} 17AHlY_ɛѤFr-#¤`Zr`h"CpTA¦]yָ ĩHK E[ {H' a-v QmXk"GYb0;mˋ !i<{\ט%iSC gQ!x&-h,o gV^*wˁoӌy{>5mSPTr)U!u|. *%0^G4s:(,6g-hl VL(.heϻaa @'7XUZʈr0eg*zkxR+PQ2 /HDtBT֖)0_&;X :<,(ިO7z5]Q"Wm]@CxlDo\ӽGF.+~fE(^ :W1DӁS| ܀ u\3cHׁ̓/X~;xGU(XoVHH)d4`?q糏gL+/Hs\hgy뷛HO~K,71dK DŽGNzyDG㈹gK1&H|,XV,zP/(jFg)|mC[i[6T3ɠ$h1dkaxLߏ̃um"*nT.K TeLǃp^MeEB,W^}7 `y? 4!g"G18A`cӳ?2ˢa@>`L&v2"1ȃLE%+Gqj=vUe\̮Vz/V3{nIHH0ɰ&!Z3dqٮzr}4WKrFJ''$VcD ts1OEFtQt韣M#` q i͠2Ēq?0'(YuSsp ||t69F^W0Ԫ@$0,pdsy#A5@WCg kR&]SRgmХ5fɕ/a؜|O;?޲_4̏rOxv),(r5α"QdrԌ!rV+/+J )틠#ʑ79U^H3ˆOQjMMek9=$SEUVtd Y FctK斌jzW8\fM^N!g&6xBGG.ЋltJO%1b9Ɂ"5HzJr*pPkW ʓ$S]C,l)1 .Mu>$ޠ4&cZ%OIA9b9+ltՃ ]c2k (Xa ch%C9l:EYm(Hf {^\^(,̪sNɲȳ LVVbAΈp!xK(?ZdJ'r(K=L^("ΈEu ޶= 2f(hņd˲H|YKXUg1u2Ege;_%2YPFz=?&6 7'e'0:0S[9P \`HnҴ=&1eI{L.Q3^YDsM١8 !㾠 J ǀ(1"e e'͟ -ΉjZ9D(զn.dYh/nCFydJb=ZOnB ZNך [2FJ/$Z粕%<_MeX˪ G[qHFO8_3 1~貈$x}Ǔ"^Mξ4| *dU:}p+ΐ qKxd̷:no-Rz3-T:WK\b-AE_h׹h*۠lμS棉sv3磳;Gyk`Vp} cx[ZxOdp;1y]hmw y\23-e >gkG1i^7NzY|g^@:*8XWʴxMG}H$ÈgJZߐ6}Kx4驝4UB9tk[2 n֛0y"ؼ!#A]C<}8Dvc ku˚V_ ޤT6ٌy %cA}{jޤQw N>v]K w{ˆ,ktOEYmL2-:JLx6JTYxдqVCLC^2Ԉt6.ąrg_pMX*'. Xvᴍ ƅ-o[q[wwzHܱZ߫[o#'/X.VL?+꘻d:S :WyzϸgKw?Jjg߯4L-w8QJ)Czj/uVn'Zի́Z}\Hdy4;cW"=Vh?y8頜,YYENCB=3.X XY9#jސ7xz8k˜uE7w\i6_7Nc~T[3;:2NF^lhsUAU@PAm߬Βs^?ʌ 2 [@64螻kq4-{"LNomM" jlmx-o(8Ԗ^JA1ܧE6^e봉o&1ߖp|:? 1=zҖ`H}]^[rFWߜIc37a\Z" {'gw!?oNKsrRt{N;Lvw7/X?4pZK V[(Y}q#EZԿ3W{i׻^u{zzX;ϛiJӯ^2!^^-#{['91e9IubΥ+Č-g:1od6H>v?]Bp4~`j3R}#5`@QToֆ:D@miFMT8-*嵺=*ʅ}Jh)SQJ4uD9r:Mj249 wuqP nFr]uZQ[3F y_Ų^7oƼ) zU6_Wqon_7ޛcD'-wŽf->_ TW1ע49˖ʅT\S)8R-UӢ碜fJr԰aMYYcnh{NxAAHb.8P쨙2M`ecAgVE xSlE%W-G^ #JnL2EJǐS6p##sD{UCә$dJ:ђ-Lf3KnB#fFmpwB|-ƻ;1Zē\Iv$F)2pX-1ga'AA(6ZC!:&) rpԡk2$ 5]A_zs ?ihz^Z |lj@mBz/7gl:uZ G]|ifZx AZ#fS$+Pg.DE]$PAhG*@DĨ[!"` Z#ԉbg@v6(3}7EB?/*XzՈ]CЉqS$ ) NQu:SZ8$xT/\MuIHH}WN ދ00R"Mea%1) U @kX0/G%( u\F0.QZbL]Q1TȨ1BF:ּJZld dHqlJ()ʃ7og0ި% bnqwM'p$I(irHaA r]9b0[!JqYeobǘZ8e RFIuޖŸ.'R+=%[+ a: 3UB@B*jS9m8,&ApœBc.voko,\fL}W}l$t ]mj ( vv"$駼N~ o>hcXϘ\cQم#$8B_Hu(D I0&")BSY\0y8zHc}ީ!^ߏGdgx޵HAP~ugW9"^+hD7zm|8*#ʎvszјSp8qZǝg.J"S oxb8Z%!فR-s!kh)[*uz1fvF0'UT[_@ *ahG{.G qLRoM5}IGo6Pޓz,v{{Y8}j$쉙=_:5 O}Gg3PU%Y1KBv u fEWBv@D%͂JX!1WYpa 7p!t|lw)IJ.nr?4`<ʼZ;[%@cv (¾7m4ljMw.f.΢So>A TcӦ?MΘKLW']f[+5tGktqh3~Y{*m׌޽(}rgˇ؅C^3,97Arw]U CIǢozeG$_Q'{%]d-}WY1%.$^ȿ>Vݟnߒecxc0~ZF 6lDGzqTNyyĶ0Q:iWf `z%rX~82c֭R׬Nxֳ wͨ1{ŅwG<J(kc,MM!K)Tul՚| ]ne} F(jPIf͑1Qewk8:-!{գJe,5)uqbMfUba ~sD?:ܜ[E- m^J1w.Xb/6B)"SQ G :Ed cn tj"7Z(ή m c {I۲G 6\m4Z GVdd1S"U–S͘ Hjh:ZT[dz%&}ζ`lX5P) )r}-O иCI"L6{Gje>T`s5;.ZPq2Yr0uލE4RSTo&_Z3fgwd ǔ81eyLyR!{爑d*nQX(<pLgoqyJ Ͻ5| H1t㡬5k@9hn3q&A}:ҁ/ɻ:aZ.HUmlR"\u,Hl Bil4$ fu6:k;)VoXKfm tgIUL6 s&i]y"= yFg`ݝr~u6ҽl3f/BZvOG2׳ϟ ~f X|O_]:x I[3RhV6Fmp!Cw{޼jl?g7J Q.T I䬂+V]<Ф]6M`:fSPoun#qK.W\9QlG/a@t!ufcB)Hv3@xCkϩbH:gbt-\{: l{}U95(P,1ڠ>@wⴃ̍kA JA$ S.R@$ hk}dd-k >g;&ѮNX'.B< [1~ՙŬ[[#L+IeҕDͧIGSz24#$6k]jt> 47Bn0,98s>uasa{(BqYb|uGja(ºQ)kڈ1Ɣ'9a`7k:Cik `!6[dB,erCnL?BJo4Oa8@ekl'JNyRkMJΒi~0dCP:=.lx+R 6M*zLyRFʊ k}mZBnG襊ik565?{׍JłYY7q39]oVh_*SBIV\0_V^;KSwY {^џML|.;{|.qO: !>|r??~mVݣm+l}fkĶXTFa(F[1/OV@0IU 1@x/mܠ4o7}z C ^(Ny/o |{İ_-y2  OBC'eׂ+F])FOrf qf&֥y(qg~fK }yHcPR Wih~g9ߒ=wWK/2aɱqRFguׂL!QTTb\:3;κB>[]O)胦YK/ԫ%OƥbK@CR/Xz-ȥgR.UB8m>FGt]-y2:'&]eL1;cuZ K@ `EL[#16!<9q 0Fam9h(/L,նIK g#Eg|azfw>!lu/]e@ _<)ET) oghگA\Huz+5}Uk3F3UjɓQ-’Ġ"cUZ 08i>D&c:߾hRQT ˂0 :S3]oHHxl/ݟ>E~ϥւC=}s=$`0-䧉Bk1.< !rO! <(Za;:_lBm? 76°%O%Z︠:B"=VɮY" 2E?f&׶B#5A6ׯ {55y䯖< o=i68X%AW\1# k4ȴ3c}*s؋N."8sr$yĠs[5[zW"Ւć`m!BS"'G+ANNNi+7{67۱=}7He|&h$f(nv}K,ҩ°Bqސ'@#%/i!2 bxf 2o2j֤s2IM݇:,9Ēfnw+fO(dW"x7S@p"O(^{J&çHnK_ #YdO?-%5޽\۩ʵ&_޿@ݝH\!{brݘC` G!\~&=QJ!DqH m FsB$VYu#[/|?i l,?s۝ @PԜiϮ{t1md'yV|gEͱA W Ta2j<ۻjz?26u@Ů$ᴤ;vng/x-Bm8Bرmki.VXBѥ<8Q9d2o xE5uN6ȧu!x!۽?x1Udw=[kиU>Űk45+gl/!Pi"?4"4ǔGқVҲl nAeXd?R`0mkyNp_1腀 !7SuPo=h⽶[1SNaHOTL)4@̬M 4vC#*q iĥ:G~]*\.'+WK[I%jAjD;cʵ o? L+Z'8( ;|U3#u j:nZ[PL.!5ZZf:Fɶv&YDO$;۾==ul6^s1|tEki'gG#3Aq9Otf;9{L8 ֙F3d}o(rlݵ|_GQ3vtدki13@}ť] }O\ "f SJbc]rhTsDS<(&bedk$XbK EZo&]W v5ك~ Zn/[ _.= 0QIQlY8J$nt~tbG$q8l- V.^~ÈwV.\bUݸ*M^.z,)Uľ[~;vWS6߆;On[CL)zwo~I"4'tKϴ?+`%̲,<}W>tB/hQ3aô/uctf#KogqݶMI*S;݇>T?gc9_{-Yn[lYdK┫"Ql@?|stf $Kkv޸m)=nOmM]f+ +}fb֩pwߛ{]ēgm~M'꛲JWV`(M&|HXHWFCˋFd22&i4:UV~)bEsT!+j8qYMbR~#(YP:䐼?̌ƜXCrN؜.p0OtӲO.4-;7&;RFy+8TBLG v]['Plɣ_*P!su@5jFȈ@#Y=BDvrUV\LDdF6gOGF$~ R_߾y6b1|.C `WoEx.,곟O)>괓 ?l pt Rpt?D헿V3F:ZJ<&,MN3aAeYSX]416s׬(=*R̺Y4.zٱ/U 1t5]f]\h^I<^* Z ơ c0tN؜^q.v> %Q{#ZY1""X"! 9D15#"Py>~[WȄB)PE+mt"6cdN m `Ym٧8 ; R UWrl8YTAJwpm88XvlѲc4n۸Y޴\߯<PoiUVP%^9_ښHIY58'(fS9ct򦽔!aGȂI~{G)dţQ֣H~~-;?#Yݎ+1?9eH~I|>doNЈI1iU0E wGEZsG͝77=mU0r lW>4(spa%ddN`iGkTXU,47"%zI$|9-2)qIBX]5*RJ,f< ENꎐsz,pzM>]zg=N(2` 'DҸb[=Vr/3`o:xA,'[h<@쎌sfN] |bR|ٝP|BFCWo|DvJ%s%^lH2`J$S4VW}XF*L} =Psևâɷ]v&>]@8W8~аiՔrgXXPTV+ȪkViXiT TAa^9Jbbp]`%)[fհ +P1bR1&AҲ/???> bU6b˰FhPFW*D$6I嚱FPH8hI>L-Ԁ+6hƇRsj4 SM7P ql3ādX3UT:%6<]:Gj)=wH-n'VT11Ǽ̸LŢЂ7@ikN)+ نasqxƈFVfS]h_۹3dORSPEH~&h4L!#s'bdc7.M]:Γkɑ:)ӯXĝԎliGڑvd;L֢71E_HPLe ރ ^)$x,@ ESP4h4.Gn C2Xv l0ƚݒyDQ("6U+~%阵 Waet VԾ(0L"}1`bX=j魒ҋtu9sy9¨ U%sqcǝ[Ύ5jJ}S6UP«/ ;Wn-T1Zyvpjj~^qJ^ղc:}ܗj«yFH_ vt{K&hת9L E5Z %hov9VJ+d/ *C.|5^ vݩ5-dg*FqPl)\IH.!~ qd'lt  y%h]|'hBIC6 Йz YX2{n$kɎc:AT!YKTl*z(ztNj/Eχ*KEס*^y$̣P*)d%(u Vy$ld|f,鼐B:+yL6eڜ<&s_SxkS÷+):ձ{)Ua&D U*r 0ҩMOB1fRKBXJ>ѲccH(:bMus*J/څz.u`|@B ؘT$A~ʱMw-+r‡׿ >y21Mj9oNSZca8+D(\Ǟt˹q ĦD}}a*+2qGq 2.dZأo^3_ȿć+UX@LePq~~D/~K7*3X5(漩J@tY3p0k3#2YlٷcYBY99(.i[s+G, 24fS̀9e=ψsir E%f7WMak:(סSP5V7,YUfy^ .3X+.,G 3n^9GWyNX b)OW7EpWcQg=aK׍^R06U7e{j@ywla|,Rddž\a r\kZClK(Zř(%,{,ւ[nblKYt|b^/-i~z}-ue;7+_TWtd^~PniL[L-ї 0t}1uGMM^䣪HN׫˳H{?GGKqH5 M B|:|Ю7FOL@Wh /:*̶)@X՘a"y[T܏ӻ[@8q%ۛ"hѦ ||U=a*,v<@qJor ghX?ͻ_g_WyΒﭡ=oN~*ίONNFlBkA {9 @jJ2kn97{C5RC.䦮P1H(kC5)5{sޛpsCCF]䪰ei\Yktr!iD>\6se*[s<67;Sn_>k 3['M7dFܫιz7vy5.nCY_܁ G;Mn'zU:4x;;bN琭$=]b )v|Ԝ]Onfϟ|},;RUy>YUs9#ߧh#kk%H!y)ֱsX;BpSe^rbh~c?\[;iK9AɰlaSwϜ{r U|>$ҶcbmL:VĹncvt:}Q>ɁNChwB q*x!3PQE U b93rY|NMDd,Z5_uvkt]X9wEǏ7vuz"ׯFtM Do/ X!3Y[5̈LpOgǧgI]Ysc+,ޛ{FX4v\J.;yu .!EJ8!){$ݍF׍Vݲu(Jx?W})O#4^`T0|VƉ﮶t؛=`~+ 4=@g{۸0M:=a&fŃoq֊Udf<"Fu<)3<2?v>awD,2&I1#3@i);@ 0U®BvpzPiZeb`maD￾D~|~ڤ:N~qjf7|/7y7'|Fpζ8=x"ꀚxcxj~/eώM0C~h&vmn(S-dWa)2OoraT}VpϧVQx:jij5]۩ yg'Z-{tٸqA7hFZeG Tr4_x %p̯PbZ(SiYԤaĴEUGW7q" oop+~lܧ.ƕ4H)>xbN8O s* 桱Q k >MQs,Š5"h4DHA" Sb`{ȫqʻ"_MZ~͸wsiSvqQӰrOB/GCh<渧㒨P=ڒn QD,2t'F0x=窲Ъ(2%o5X`4-gk^Q^kfE4g*Y7K(EMSu*rL:oځFnԾz0_΄xٱ7q1X_7dfnnoRzߙ٘QNL'=UY&4 vY볃 FZmop!P&$6FҀzGf qf3-'HH[+wHb&ɇbB<\F0}u8Q>Xj *yDU_V_9%66^Vrp;i'l(&ѵJ6?s&OngLo|'=zW(٤ v#zb*q&^9|7 xD޼2)H`iOP;⃲xV(OUT|miZ)E?\TSoO3o gg?=o="<⚇|usA3c.ybg!;+H9<`KƸrJdk_[d5PAzbsj/ם7/䳧?[ 4{?Ļ9q39̱䗥L㢆J?/(37P#Ûwi T? ~m X—gW}@ADv`>X涭)^\^,s[_V\gziaFjA7W %.Yhd#|htJȨAсZVZmT߬.@GD5+kAהX}P¦c*x!;" ~hHӍfW-g$pK8HGRfv-xNP6zۍ_6GN`7!6/]*@?&kX $; Rs8ޥ7~LtgN~xu&r9'KVo%V% 5:E.^Rf̪QTBj娾_nFBԐiqPyH55!:=|fsSɓO;mOOc'ǿ|7{Wl#N, hqcK tkwg5N~H>؈nh`oJ.}jiNuY,&Y#9SzuyIzG_zї\?/zF78uv/hrOÜޚOV٬inF%x)^cMz \`M)ECf]b M$q'½HY.}[-fm^0[n89U{H-b`4Q_Y?w#=P|]a6(-0-77~!+'uFhcQv}s+6Mg82#v2 uob F[8d -(ynEqCҋq kkRWX/};?hKA!n A 2I;CףbmKW,靰SsZII I5DX$$K=$T6ThDJ[^ v"dpejvҽ&U$I^ ^8I:(2"BB2gKjraWTFLUTUIPU"\^,HVQzK, toDLE'"GiF E ߙ M" խ%{URc$ዽXCmD^,*[EًZ9  K7;|`XuZpR4m^$p]9iV]ү%_7̌2bo*gWeT P .[;ENT6HÌ&٠W.x0HU)ڠCT5h(50X8[]oƲWi SC@>iI[\4,l' YR(/INZEfvfvfvF4P溈.qa ku\X"ou10&жe@b&" w0ĒDW868ZF@fk̜}Qu_^%DOa*R{XΕ*.X8*.7>j尌K,u_lz/R ohg sfJ4XCxgĄcc [^Fo_yzx}V䴗=$^Ýf"ښ,| c넃B9мwu)6[2p0q~1FYwxGFޫ,/ѧ +;vU:SGrsQΩS;F1lXfKoF: F iJ#X%| %c 7p9Y@`@RHh`[m(Kƾwqh$!QDHHsJ"r\ȏ'ƖXDA59%o.WEdq3Aγ?h<V mjϮ9~FqpuwH9`n`MT@"ZFGq)b+\J0! F 0=7b B 'V1,1סƆ؏CJĤСO^H"I-Uޔ*A(IUy# Jq#5R )0}!9E|(ؐ0L7a",FNT?uCuXu_CjA|+j O [-MW7Oz%<Dhax};S0r-EƂgjsVݬLP[=1y+O|Lzj[%H H*)juhv_ͥF,G,'Iz2tIfm|+!T}Jjw!HbUOZ:I\;iE.ݺd!'8GZ+iziK ~ ՁRt|IC )# % 2ICn"}& J"aJXmym*kl~!w0K*[ch XuICnU`z ]kQ뒆*aǴȧA?b?GR2nGՙ%!0 *:e]ev+~[Yc9#vORBvk *k[C* ֭&;yiH5Djr|H;qAZ g90Jj5+;1֗|msе. k]P{)%k]J*H׺`̫9ʕ,5vyˌ1>CRV̧I8B3L853B44cZ! `Yבo` V$]^Q\Ԡ7]S@ NtvKu$Kπ;z9sr(e$:"c"3~֋v\U!;X$=Y]N:kqPr8@B!##\{G 3ջxF4$&4H(A(Lb?!6$D(✁"8/C` bP'_SBQuJ.i [-_x0(л7u U\) T$w!rbvȠV6(ף.XFk LFتd?/(a+IpEZk$JY}Ľ7f1=ܦr5~$8?5;ͱ2K 'F; ۰o dơ?) W+v7l@Hl sqeVDg ß͒ѕ=`b diIC\U y<嬀+s=$8IhsMY˩NtoJKRQ_F<%P}f!8\œYb7,5Ozt|0u'gx u+ŊW4y)5fm=0d*FՇ/lHvc03MϖyV>QvF)aGP"B!.Sȣ+']r*RRgyԑ Ww!ڑ$)@-U$Pu~Zrڪ.+7J L+iRg*YlwYI6ѻ ,f,9lr/ nX:8IPa{;MH>;dsi?,f9,8cFAi>KF0̊dٙEON/g~r-*},? R'ty^¿Sn }LξL껗\#P߳ן ʾzzzY `o};S+^?[p v:2%ûo&c:Kne#XGVƳax"ŗkA\_'fgRhb_OR^]jեw[3JF;b\+=_Iytx$q3 Fi9wӗ/,^O—&e2jb1fu=z:midWwJp#9B)G8!߫NB::_GY_G|u:.[|.RQé;a^lF 59lJe}e;>/ZJI|^rLC.ƈP%_ﻑ`\ac}^1A1>A^Uܻڠw',-rǙ|6Lzrzbd~`ӎtRaÑtq)pi~|V?=:"m(߽{Wnx(z)*H;*wE-Ƈr?"*PCD[fE(E1]uQ]Q@g3q%NjJhifSIBC3ʼnۑxgeI&i#id_cs-X=Ŭr7&r$ʤ٠5XϟJAQ}=<Ψ#2;tF}gwFV@y~[l"2DD1Xe*2%2ʖtMhGFo¬ذr0OPZau/TNK5n4c#ViՂgFRM-OP)D@'oz+Ty8V@N'plrcנ ìn9Ϣ0Zg A5ñf5s((8fAFF^h]f7]13.ubbtwwڣU&n7#dYsg 5 o?=ߎC5k `5^fbŒ8Ύ e^FC(Azx}prk$dA¥2J{'VcQ*!*zA'7'ЂNN7oSC\#YRet8"1ѱror)`Z,du껴mb(=I%ؼ^5}IjO^V!Q>GS*>6A]68}[Z{M@7eJ V]|ޕu0RKTVZַU[vމhͻY NQs^{KPP{Z]ƥw%,M[`)0'(_OcFE!-W"vyUacc񲶫68v(ɍ{z);,cW W'jG\" U\ֶC6q>?4/Up8nf @gWAOpEA~kj/c_<[{h܀ ^JԶC6<z ,_x4'iE.| &xx@8M(*tLdQ(@HHFH%Q؄!Y( /&;+<@_Y^./ޤO{ 1X0k(;6jk7`r/M4&QWeei#ƒK˶m$r?}_ddm%ymPL i]~Ț'F9: 8"au JI0nHgbF>ʙ:UQJJsA f81KrPYc>L]#>yy#UW}>U^}@KJn>\pvc Ee(,PK@h?ꓥ0\w'*ncU&~n+3&Ev{_c:q1j` r.V(FK$?xݖnI/nI/[|S3 M)&~䰄XRQj @Ú "1RA|Zκ/1Ŝi|2Xl(ɗ'ˡpR'6mlpH* @6J`)NJ1k+mH;Sȫ=,la l؞4t:|}#IJQX$u,ݍT*f~q}GVKAG =l(_sS3->h+y5x cc7{AeFӧ}؈?l6{měkӺS\ƱY#Ѫ^Ƒ 4Fd^F| 6fY噻RKфB]BrJ`EK9%-2hc>Q=bM/ @0MQP 4ǃXRRJ$*gTdteC0?+}KÞ?J?U2dtQ9S$*zm$>(.;aK[/:X3Sy؟f_k1n/GW'c(фw~W8:uix7'Ы;e'߭-cuǎVc!:C1k[-?>3=,n^ҍ槟K7ڥ~֬$vQ7Ub~TLL.ݼ>1<ܴ%t灓vQzI4Vk*v4/u 3dfKaܮJo6sVw)9KmŚ Gɮ!󟿈dP+fzPo{ٻE4J"gT,ۀ )2h)_BCZ;([pm7_W{S۵w[{W7+2k^ݼR*)-n8>voDIpk>MQZ/, {llYTA?{X|)4b^,?3lKh } B=_q(VE䜛[ (v6RIN3=?i^PC_iGgz(I=ԏ?t(I=JRן&^qMԤ:]BAf25TJhl>Y'&jR9Ԥl[<E]es⯔siD]Q,z,_;lmҢ5-PӢ5-0ͷFeYMmёO_B@}䇒xR.:޳BCҺs4\k$'qbf1_MpaK^Ϝٮ?s1s~elSyIIӣ?!Ԏv2.$a=Cye+GȿZhz+=-vF<ƈtr)5hu@Nсj!*r@NK[ly&gFFpM(H2&YB\L*xI&3V\T ug.sC[u:^&J! sO ؄ 88ѶWȩV(}ⰨPog=.uYlV~Wʖ=h0shSijS߿+w8.9!/.~=o/t7H^M?™ ݎ[ Sl=x/~x'iџGw|uOo!23ƪ]lGyTn=1تP|OC0יJyz`\輖*g5-oL $,Uʝb"_/5Y&Tsk͔^Jaziz釒oe_DfH#73M7@A9-T5y6t%s OqgY Yi;~89W5uČxQ"&Jw4!^4L& C-5%S\J ؠdNs ke)T/VG٣?Ýx ~ד`la̬j~SR#L7 +#j%8DC V保Xl, IJcO a]c? l.To邧H> w@KlE@AFrJd8B Ю hQ+pC |J:deQI ,jG9"GzJRMPxK: 'gB9DG7`^yTucbUɐB Ld("EI e ׮U-yN/&E\S{A[-rFI!e7(HT#*9IK)Kܒajeq7 ls0CXQt2 AT,5>Xn(fXTT (HH"j"(_Hshs)cu< ෤FJS>_-J%-ݓ~(C { 7,ߣS 1xi TИFםt$#OQVY٠ٿ18XRYDځ%Wd^V#zl+;f `TEF!ʢ+d HPϳF.I:B 3FvZ'ZthgCVs"-eZ *6rĆBz&OJ]]LD)Ph(ѥ J(m2X җR'MG͘c%Ns)˜c {ksNB8-La8w bmŝuҨEw6Xyֻ^(=ڳ"י*X#-C,2IL$3%pF*u8TJJECqs8;Q{(ڎ5=WISms"0 *c=TJdITaӓ.,DNO pO,a_+YZ[&p<;!OPs`jR㉭Hǚr{ rcpVX㺞=783lo©jLmy]9G\Qӎڮ{To:$>Kr/u˪F֠=TO)`ZĹ\ٿZAG.JۦzFEmJYCTyS{y-D0,UXw&0'7P=Kcqecc6G Jp%,gr|GAsVtb'GC֏mIY jWaw` by%޲LU1LjCwj0jϘrz썌HCئ]?_&jO vOFF0|6;3v!-IVf =dia,:1:;kq<1Ũ@ջn^fve$d`]}4@M*,[ :YSԺtk%D, B4ϰd$}PԜ'KƛUƺX5= 'bȚ١T'!*+000`h3 3׭K6`PɎ4⚅+C n;AE66ec=!0]"ۺ?p\;cW01!&S`TB&6[hAZփjv%dKuJ9\P_ʨw#da:䨍(L=&ض+1휿uW*hm;298 ֤le4 z(Լ=-WDA$YHl 9mx]U0N`0 !tlb3nkbGi;p5˯cs ~6,m{''VJ5>ʇV4FD ]2rT|c稤/obaBh$g GIZE v%*YNԫT˘bzUb~KQީH0(JF초ID?ʿ҇ݤB}8.Yκ6vTV|ظT=}Xp@ bfz fݯ_(xcqAZXCȢ5'Id"0A\T+5%ixT11x]s4_go-?;9ou4E S1V:rږ\q!R!3 J]r*>9 b^R0xG+jE#EHrOd)x|ѶTwsݑio~hU7:\cR gC8'x"Đl%#Z v~s2)36^`Eޕ%HfdA؛MQ|S;ͯ=ݬ=FЏێx9'}wo~6#8͜_x?Z4׍O! O1h$D K=D 3X3;-AWo"#RX^ڧ\4Їow׳Bt#,m)n܍@o!ƟzCh7ilf=z1184U9t&z!9Ρ0JU+'2^<6T)vYgogՕWL;hDPS[ `iS䪤M+Wp뮒Jc&5ISn[ R!}M2I K[3ܽtn!NϊvĶ=]% dO~uhu{7Y U]~0YiEK0^ވ%>hT(אѐ Ƅ)%:3-;j^YF-[n;5fl#?!zԂq뀋AM~78U;˘bUhR^KY$O$g%xR̐ ;cWJN){A lvZ]SX'U6) cmO{+LDX1d]"v)Yë)Ç6&bm˹7ތ˛ѧ^8!'qg/.?|]n}IP)p}>*p6AZ(y0m&uvMکNu̘dl:R@|Y7a%8]UO|`mu3&%XxTfDm^Z_)LȅJWD3F9&_TPG/*||QZd. LiEُ?E*>s&9+$WeIrGZθ1jU\_*QnHS[ʩD{b9mU4BŮRj짔&4tj&"HX^ݠNSxVa<΋n[VZ|[MlnC:p&DG@ʧ0J^h ήNZRʇ:mE%luwnYL<e.82'.hJUd()fy)Ny2 P.\c}CS: 9ԌkM/]ea<B凑:du:Z}WGq$䬊+$}UHGqj;mŽkpJ1e(6$J8KV\- ˨)PsLu7NiH5iuOC5#OE5ZcYjAnNn4":Ϩc݆Ewmݒjڐ߹&ɔgI;npsn4":ϨcxBoٛuKFukCB~"X4e+#=Z^NJEE$}w[`#0UBh%QiF\H :?HEC&j}(jҼʊhjaDSS1SSLM1J\MT~~ rz` 1RgQOR3Y&T `}v#B5\i Vq ʬq <%&Gҩp4 v<Kq%ZI[s C,pK,QF}.9J,c芻rUH[7̵acfF;mس㷵?LgsB;f)՘ !L X9I T1T|wv627z jU~L/% (4m2,u!.:#qgRLiB K Ck]JKROP’dX$PȒQ(a GU-)=U-lX)--ʖ6%1Ԍ)P%gI˚R)mjDaKC2 1bK+옒#=`-"1l xt 4krb(FW3&wC0#Ra$<91R[HPƼ1!%+Hx j 0>FQJ$3B(M(t1XMUlq{.|~TM#SƤFAv (s^ XLfW>c0*2٦b Oƌg+oo~'vzI-'*iiN!cre5:NB'OKV${q'3Yh g\ 2k8lK` K8yTg81)OF&>JBۍ*)D4%(D)g}KbV2!& bmG0a40cS(BJ$F M)aggI\cl`:j%,OW5,$zJrPF|lRջ -v‘urM$DBOSE!<䰅ݭ`|jsG !>a1[^ 'HV=bMb'/7(ׯ,'O!u_/ޯ7qcɩ`7.?k@juc_]Ÿ_,u33͈ Fds./}v} P z]y7`~YՆAC00EvU{NFERb%N­&u;QI2 IbiSX-.X 1?,޹|imXGvzݽ Yco$91|~ö ۪]~8Ѫa[β-PUܕ`,] VԑOfmo}|Zm?lXͬ&d'7T9GTь[(hTPFL6Kʇ-R4m2F|{m޸hR s.:,OoW旈 8qt PU0ONa z8{אaU~<{4fXP1X#1(_4 ^Dx9l+so}v5mn~|_k?]|sG1xKQj^|r~nww/]|9sW ^Woin A[K4$Wxal+=+ kpHa4X& fP(nK#™rLp|Slz{ [W[K/W7`m%ZL#29 X6W7S=Dk! bsٟO8f ,L+y])m7啙_l>{(_KMA/>n |׾ݥ u!pS{mGTGW H,#X}Bp^"3'f$00|4655K˘> J='6jTF7rߧ! P s|75uYjϐ>᧔ؔ7ά@Kb3yF)"ܓ9#rXʘuA$"O{:|ONYlAzhؠJN{ըnMc ^Ɖ1di ~)L@Ą $AR!́}Z\tZ)L\4+꠽d$Htf3gq'c]b +LƔGsԖ*WSa͟YXS-9a(,B|C^_nэY oDg/.?%ܽ;̋|wq}/4A%JEBqYf@JՆ((+8@=>%1a-OG$¿dn,H&C k`g2) v3c&lG2* (b)jάLq6 T )0 FHNH#8͡&BV-ɰ pc6^hBC#D/00Œ$KәɌ#tj\1's\j.ʻMExb9wn7% Iئrsu{Y88xws`K<, ½#|LM7 _6tԂM<^s ^vnfWG m5ַkmI_E0w;qE fƘdw"QJݯ$aI0l"U;Q?f'^oVA7nI0ٻJ )uEq`ta%j:" 2dCy_BZQH.FF؂-(AT{E>(GsHzYv֭ULZs*Z: 6޼̓h^^^#ӥ)TNf\ 0&Cꙺznn>'+\>cux{ayhh~r~TzTnث'T5TU&IUDq֥7CKBQIBIIQȋtAhr:蠵(*Z GSM0AȊK"O5ẘSڕ n(^SSͥ@Gbij`p@&L@K% (Ɨ"TDZ #nD@)CZc'ovo'xRRˏgHm72..d6x;g]exm:g]b೉c3hjcmo_gS:>;5P<]^rygfļL8=uAթFvFdZXiۺfukBCrS &ٵn.XQmTng3A[mFZ&4+WZ:h.vCi] ʨN6X3Gj3Zֺ5!_ҩBqZ7VH?uAթFvyk'n-kݚА\Ekf_ ٵn.XQmTn\bRYڌnMhW[:EK WzX`0s YVIY0q}>}q",/o%FI{= J!ErH1rAR R BrR#EȊR8 G34mšܟ%_G?:2MIˆĤx}݈Qy0˽.9M5f&HbuA3GSCD KLς4Q|ufOjaJK]dtm禁?xgq|yUw~ߌݝY7'vyngz"=DK]GCܼ}}薟|e}g/y  ]0Dz " e\5Gh1l% @9%Oο{_,a>[Уf?o{H{E m 5נnO{]p sS@V4h\տ߻ype+3]+YLwd4ӌ!9a EՃz!9C#-t2$bT,hbDK CfR`V>uLJj6vy%N.Vv =aT%Q0.TT싨)Y8S3Of!Ri,۾߂5v4ķ;[̮@?r3>"2 du/yעG ێh +Q$AoCnu!c X;4 9 Gц#kMȆ=RzB[Auq}fJl>j>zu 1%+b'M hu[qe Pj,ĨrʰH:/(dh ƊRAdUE8DCމR\HG j§3礌(kf&Pʝl(!)x׃*7̡yU?F+l,Hl+L8T$4ƿ7ߵVrmq(P p(ID41yDBlfdXJʆ{xʷmbFPXe\P-KÇD+uɩXXpL;E/i~^,ZB,IJ֘ QҋsTTO~"$Ke{4pK>~r7~M^#^4[bp !ခ<0a-*R̊Guc S܎UuIE.\NZ>AvPs7Q`!L E?_sl(eɇOF9„p4Б2V ]{ʡͷoϦ_SaC!"c(?h qd867Zs*&e%G/~9fI10k'Qؔ.6{;-_1otfs:K#wwv,x'6@ϋF>A^m`wW:piܞj:w9{.(\m}J])EM){j6E#;+>tm<(yuoRanl'&#MuZ_g 4> \t^d]+D'!f) MuZIbgn*172?<:l9iːs,O챙x4Lt˪H g*:, Gl~ {quyKD ;?OtxKBN&"gػ/nGGк}d )U`aG4g䒆}w;Q)N>އcVŦ,cЧV:-Z66vtp'`7*v&}o&f>EM, [p:Aiz{Bn<0:=O6^;/B+lhTK *.kj(Nnz3uUg}tgHWwBݥ ҡ[QK){GSkX!UI#V3k 궑Q/KK遞D +>Oaeaխv+GE,mFZl lEVuʹ'8Fd-ʼ8W3NdJWfݨxL_y7e.xӡ?iY8=XzWOpwnٲdI‡+hf~j`vgln-)U3҂yYx9son@KDpފF\dq1A &}$A{4JiLQ{QmYtLwy=@&a%;ض1OR~f8޼ gq'+ v 3 i/3- ODziµIr{{<RO{5Pc xǾ?mc+>T"4AIaP|8 )N~wIYZJܕ᫶H#";;3;~$=]R{E Z$jl}:Jg$))%7#fP [s<$e_7ώGYQvα˼Nhm 8vTb(Gaׄ A&N?MH8)X6} [OCiɿnUIY՟W,Ӊ5NcBRhլ qeM؁a$7edQa:n]v"5;b|p&1JL}beǏ4qirĹP>ȇG='LmMd`PwwQԁeEwihh%{C}*(nP*@}2$s[޵YZ3x SN#D=E"|?SDt*q]UU DLt-4[TfثO}J{c6BgiKo˧ "0w^VKM=Im{f hSNxY_옵Jݔ5 (TB=s[mq8iOZڙ:dJ^eZ;эJn}9.˺fZ0(Q((Yw׻ ?h]%Or?L5a~.?1L૷4K,ÿ 3" uՐT2 _(IJ"Uō^ 32񴾑i~goʗ*'O+]UaPzw|м#V O)\F^@ |;Dq.Gd3ju3_u6X-<_WD C 3DZcغ;x/Kɖ.$sԦObS]ӾlfNo.NkM@%HQWEⷀh"=E1FO q 0j fT S$q3.Q/I<0Kg'pNqp!b9JgȘ0"4SLC5&@f3XL9r~ړ -ҋKZ& G BxJ:HFK[{$HJ7ٙAb^ݛwG!u췓uY;O_k[w} LR~B>W/TZƕ=\݅xHćA0(yX$I*j̈yB3@kvj= !v\ZMEf `,.aw&pf JXN) ֙|a(8vPȣɤ-e;zGƶר!0B۹|{gSFVR$_;'erh /Y\Sb;O(o÷.vKc :N~,Q?Y$YxaB"WaBEViv~3/芵{Xi ;G#r^`t'SK%`8鎒1 d~۷Y(6"FҎ8/u @$QvNq2V;t>5{Q[#bmx6/[GHgpF͖̃n5)gImlzM\Ғ9=p$hHzB?aazFIjΚY7A<EF@4 jFH\]@4Af轢GTrGX̺pзJ s{]}3[,Ԙi!C{-JO}efR-Q\I_mi :MAg":)3J8CuL{@W ~F"QmMՑlLC@\0(*^Ƃ&4pðRm^HEQqdԘPL jocOAj@A yyT v| m BG<#lKqE~_[toH&!eF͖9^HJ:%w/L̾ݷ'lϨytDgn" S}Л r2P"9@F1>frt5@YH0N ]jnq`]PZ!HO >2@cN=:ü]Հv*%6F̈́{_; dPhPIpWӻdJTR-cj+qG& Y?pb,;i]!ntME@tswzȄB1X g%tE* vǓs?zM r$vZ exSH=7aJ=!BKXl*/ k8XGb܈? ȉ֟LPDVJ)c=:(qhCFwl7w vbu5re|)S|Zz2|xBҚR؅ ~]#ӼW\kOjxiΦ:T$D ½=8DlKDȋ=z<* ;ގ'P>}jOeoI4b6ᜈ9ڵCpWa~ +ĚSMʵۣ;?hOB K(}1BK_ޏlN) u2y +[dD5Bꏚ _EPɢFt\Fu'0d4h|z`<,?]fq>fP 4s];Xfsq:=U- #UrC!OOU>WWNޢB(-ll| =Me_JIKt3#Ѝ~d'a]k兹yd"%LܞXNFIٹ9gzK3`gM!;nfV\-8L{f_%5գ+t&ʞ: 5WNfR/Jm^sⅣXT,3Ef_S:CD A*0"lyևO#pȸ+배ĺ۝9 ZeYm'{YuUZ[rsXOOC1Wٻ߁)Шam9{cga <-|fMhM}D5ĖH+YvĮ}}vO3sA\ܪ'|dfgByvE egnRHi+ :Ei rrD]0ٍ-mOlluH(b2*\Ko< &J?QLfuO(A{H%ҬHZT*[2Dsϣפ_nW$~n@% J[5%i2k?buW[SL8uAEQWpј0O_S@O zOoლϪ\W%-﯇ojiß~_ύF>Ek]WPJEpG4䁏}bI #34c*Pʆ~<9ci6B+CQHrb}=@j77gr5pX Z?z.]}( !8;Ҟ$Á^/\! J K'BbFyeLWu@)B e$esLb?8JiEɖTR\,AoTT ޚh1'm~pl",tRȷ{]h59dǬ4iOD= 5v*(|۰ t|s싦BJ * qΝ^l );I7|^jH'jw|WIy~1RZ7'\oS CW8 h@$#F?`̿H}5_oe5Әw9g@ ٺ;L/!v&p~gq&@o٠Ae1y˯W/T/])~)7U}KpۉjĞٟ^'hOG$RYYMT2GK̉*֎}>q0V}ONpe Aa / z8+?|g |ah+?-B?gGW1э6v_֋6_ w oq7Y;{cU_IDm=#_众vdI?%. xPlnp&L6o< ZSoGaZ܀3_l'mb/HplAɱÙؾ$CDmdu-oO*{I}d7AWv,lW)KC`.}6`UZ%b\5`ÀqOs5d||Ay\9(nt&|͕D,F1ot#&s&!48m@ Ԩ:bl\ CR?VB/$s<NDǂc`R#5ԾiEMn}J=71"(GdFgEoJ'>%c{c}{&[^A; N*ξ#ͪ뛄.xug>*)kݱִ;B9'M W&RWٻ=҈_o X͕D)رJvdv,AVD Q5]ϗ굨q|5 \Z[Qc57_K_gمȑksKJL˗#9ŨMXfJ"-WJn9#4ػmdWTzړeMLe2ٗ\ ȗ#I&[ ) h$c"5GKf8f{M_dG,wGh{s$2U;zL>-fnrc6-h[QR߽oB (f<ɵ1Z3|x#XPT2@d0>p=6r1/ɻ`w56(3M/>wJ}hܙRxWH ,xZ$o߹; {D`,x-sXSmWQ1_(ͺ?We6֜ӠIS6 a,¤ɀh0cixR=[^  6{5W**n0kAy )YiqbiIjI2hݷEZEމǜ_EI~ܪ=U `c @:QX#Yw:(vhr )tѺ7ѯSR U/mO?ړ$(`,PȆky EwEXl^VvkOvR>8ʏ/>uI|Y[DF1=`V{1>s#3*ض.0Pe>Uܑ ~A ^TuhOQDioY#n5q6R77*`Gu%aiLbBHx"H4?t3]H1"A[{@jd7#3dk;VM#*0R"F!HE*02*fbxwf|sy2Ij$řF)6V 4V6DP2fe*udY<.h~Ak$v(IRDh;T1$Aʟ Ę88"$ G, ߕaj,Rbĉ )iӔa! 3HRЊq@Π#6ȯ`K[IR2kO4{VI:@Cj.W)a )=4ONRt3>y_^<yΔٽYH$@*cdJ%"TIR)c~f~'=h{DIRY׋.d=V+;>P πhS\^ S5L9C$U8J@HНbZ b>;hn0#wJӔJO") ?@0K,9|!*ZU=s|B|8k<9FA Z6wxߴ_xSG/Bu~e}ʼtHܓLҚٔ%ajԮ&NB0gQХ$R޵LFRXc]ň}";pH1)VSO Hw?{ [MO׫}(NcUhI~cP8@,ehL!dAg1ƲŕJaJ`~7Qȶ8wBse$0mR`YmH" S.R _S%se/x}nj b8wdA8'g}1:czDijAq#);` WpUxw8~&K !:jbx7 [FkvMt p.߂!sa&q~9 .E%99 \C(%N@RwrTPV(`6Q~ ٢`P+52݁FAYM3]-ٻ ߀Q'‹tՏQ232OuqJ**_&ORLeR')dJ%L8x8Z$NhcX ii!!jK77[{a~AF׊@oFe [y)^o_d8Iߒ(0^Z%雗ج=/̕{:54. Mpq4ӱ(T` !ULVj^F,e1%hfQrP3˛<|3„<}]^CCݱ%䫋 e"Jp86 ̑L@0+WL# 3#?N&=_c0pL2- q`2bS"M$cJ$cH2_Ju7W[Qci$U  mB7ηI/wj1t)ۓe{Jm X>=J_(Aeޘ/߾]sOHX}~no+W\ۏ~waK QG]Պ,N2|J̦{-GSTgf?Ȭ_J1 Hdz<( sK=9y-;15'mW5b?3 xw񹣬)GE eeY-(#D [J4FD+F[^)ʼل({& rU?~ 1NIzm c,zÆ7ۿ*?ac"?/00he&^͋2 OD<}X\D`^i"}:%JT7Irڰ r5Ӝ {$MM"2H)8).W$C2Hb J%:3Mۼ/KR4cX+!x,R}|X8A~ՋJekxT}w`o&WoN*KeD/fu۩ylS"/,abeނLJP\d8Ke ($1CqLEZ,;G =FP4YWbhecen-%UjDR%1H&#S q̵C)EFR(LK@.+ xYOvkͻ0f_>eM  T5 Q°Zk%;TĈΒ8 $Q)]"Gޢ /" ?W+`G`'ʽCm-D}K rFP0!J9<Z&%=Aٍ_cu<Í/ =+}MJ;zT_f5wӗo$QM_̓>_F&ma5؊';̦OƬ P {q4^sŕNCyz%JuqDS8z4?;wjqo~38f ω:c䄣p0JVbGwe&oP^~oui~Kj*7>;t<(hM l~Z79m03A>ؿ'yv}D0#k>Xu*\׫=lD2DOo./P _a+N:тUҹXuSKSi_μ[t%gF%!Urc^@ d%I?$EI3\1>ujʃ_lAwA~My_uFkR,L9p!J>D#t"ˍYY7'W9u`Oq!(z=Gh2yUK; ♍G\h2 V|U<]%Qiv=*VCD+a̍}/WݩF1{\h%X<;D Q2Ggpa4CdH0]Fciu!+Cd^T} Hu9wI@>`$zL+nz;ܙA!J!b`(1-DVq:Y#bnQDQD?Qƌ1G`ݲDi# hg A?ΙG]f0&0 !NüA4N Y O͔|j_ɉ{Zt->&Lz-V}k; jX@i\/U&ݝ]mZfjv`*su)d3DFnJϭ\sj_yh~'GYU3wP%ul/ϞRe+}5ߧ7y`9yzE)tO7=)c]ƾ<}WFñ"-̵؍y-[WnߥbkBu[HO^T,:u$!_v6gm^e@5;qY߽q@,vgւ *E*QL*\b򅋨L1KB;SsLZqVȎټ+av_Oy4{nw僣t2kZWr:/ˡ/-{o.˱>ċh'n׈wyC[! QMo7uhd&%i4#ّ[w]9J%`3cjF{lXְwf/8hu`'|Ҧw8 J@8xs,ZDa u *<bf`J`&{H U &#h(C? a>yȸ R-V9QPSb$U;{1@x7*AɄ 5c>oPov cEQXy$ﺠV_${qZkbv O1QYc&~][f/S~9KSaSlMQ*)e}|kmX"z`sֆiveXҭE~/Qt ԁ%`Nѝ-):rQn2&[yB("PXBpg5(b;L^jxw-t##$z fOoqwĭї')[wo݊:E:=& F<{?ݭ}vהc0HG LҦ9ҹ(!s2sܴL"iԕGu @g`/'{elL oYJ,do"SQHŒ2|`$V_Hm)+$-_'C=EgvĖQQoȅ$5HZKZX_{/gGaaz^?X*PUS< _Yc(1b|*U]!@1/ FSPHDPH#lՎK/e(,.ils sy2*t-7v[EU~e|oVWMlJ'@bCLtж OduuʺۙXpzn1ܕf )f)Ÿ.%|Tމs0Ɗu)'RJ=!B[N]r{AgPb)@S%0p(PoK!2麩WOW%*cEGDy쵡e/KE)S!*8BSw!H)KbzY+߫0ɤ?/yoɤa64҆fSU?X˵|5x@qV,+V0zUF{^p~uϜV8IpdQse8'C9_30B |:^pP!7VP 1&X")X'e.0Lw* gDݡ7BA6loA^y?N9)0fQݔ;'zDU+-H/\ҨB A^Rɠ8Od۸#m_˞#$>x#T}ZaDi5k*SjP$T)D n)^Si/`(i)H554i l 'Zf8>BDmAt24Ɇn'hz9rAޞb9Q$uOR,R~^xN>pUګȵWk\[?]LP)DF`Sj=-&z7A(K a-p(O(tiW'5>H ,;|#[/Jh)(` }oQt"Ȓqɵҥ@ÅĶTĥvⲭSAHaX9'RjJNphkIV≖c$=gv o8QY2ۨҠezUjK9' ֌?ݍf,]1sjFIDP2k/O|0F*![SL=()߃ (9ĻB.@0TSSpi ?!ڞJǵ#`,F/^ !ˌFb%)#DjiAPM(MɝJD_WhB9kMFs&&);dsX B# BY(g@7 ~ދM,/[+v` hD5XJ0*%# ?c{$HslBaI fp t1IY0 ; |# hA0%vY D;A< N SArje 46 WJAʒNq' AcGtsqͻ㚍SXp +cB,!= _áiKi,7*"sT!X5=MO9R 2BZ7tx:[Ԣ8: [deH\RW'(hQ^P#CeP1SN!cU#Rf-Lt0A) *@6̥ohwG}NOቕ#Ī݋Pi/,XyػI!nm'k7pU'0x3r'ꀭpX>j񵁘D q"x[Nudk'fvt,"A %ҪENX I j9 Ar) U*f"C={ HGyHW5s~^,gG.z3~k='~ &h &xxY^{!Q͟?ym,lq@U0p%K{ յ77j~a_8~NI z+QarįDRT^KzaѪ.os coC$>=C&OȘMjaR7d L6;eLCh:FRzޓp粍dPP\4aDЖZ1F'[kUlxǃkkv9f9$$fjFj46@<o霝YVlxSQ#y$߶MAe#;tHNbm1#g֢ _{K4>{eTlo=A9\gO@6`8haρ+s@AWm9tis.㫿bMUoiZMX[u(Ɨ]PaDԤ;D]ݹJ]?e2FL8*\s\@:64E:=徚vyHOFsDgRܦIܼGْt8tP .7㘹 $y>F[)iS#m}5Ҧ]KZo L"%(ɬ9iHc\A-{Çe/79[pI{ϔߝ㶛S~.}o/aޜsۺ.M;?w.>p)\ ]pm&96zbZŤSka &Yۈ6Ȓtͨ'3:5 R;5NᨮVqaj;0!V_2;_J J)3vA-5|[Kg%  {Ø‹^xl㵙^Go& Gl&Ķ:u&AKUɣC<[%P$dHj\:ݑm8Z6fww"q4"wLj~6 =|N͗đyu~RaaힻڒD H'ƴJg*A}3"Z%@.F0XGfvf.ݙ_cGT u^CSt4]ŢeWhU,ZvU/ZVȵ R`FZ+$[Ě)A!d6*΋/꼘[aᖱ}Ƙ>jJ?lUK}h9Xk9oWA? e.P J%-&iŗzljECC Pp`A!Bfe 4 LQ,5uD AMh mu^%wHEꯟ[M9HYLmlRiX  b#Hy%RɃ(2_YogY[lӠz;%Dh3T)'(Hme<\XZ Bm,RR3DѪO7~ [¢Wä!G@B zK+| S)4;Lqp s*FdZ.i^=[Ѓԏdw5Id|{Cs'wP}8f`?y܏}ot?^I4ҿN_`y΄ ыY?zQeF .~$S.F(GE*r_HCg$ڙd+#}'7~gVsU|wuc~h?_M~yUV=Y\ϯYM3TZ폯zG*4(HD_˗k8az7ۻ>A`!R+ ~z]\d:_^W@On~ܘˑzhN7C [棻k&꘤8fx_?xă(`G9raj[hgEԇJy9h"Xi]kbBs[9raуWo4Q :sn}Fi v޴[1ڭ 9D0%vڭHrUS)JAȴ^He0 7Es->8H'Z|cfÓrlOSBqɓ3]ifCTA1 u4rAj E^9l*2-t{9F53bYb0M~{UySt1&Η.Ú)Ӎe%Uqi%P2:ؓZ@ L r0XR+XlDզD.|y$dR t9Qp~M)r96=ztJQ۔Ɍm@ȑCSL﹪T>҅'%#p3jdc$dž#V ^Q)q˼V*<>`<~b ݪܭF iJ=#Q:x7cpd4\P^HGIW-A*BE,[>HU:|,x.Hd2X''ri8K<1Z %Dr!fA G>XTDFg""^zM9HB(uyjE_) hUۮR%9 !Z)ޒjQ> CL$6&Ff71!G"Lm{nvM?GVʃFEh,޴[1ڭ 9DƔ#Qc}ݵI:yTew?TҦbu=k܎0./_8O<:OI='4R]r'%̇OJ=>?4$?xGf:_}$4t@$-XفH}۷ɞIX#ek; +ɞIX#msz¯JgI"HPNHֻ=X,4(EwZ^ %TAǪpmVeĢ"4A'aHzVc>X'"i4-ֲ`p,Xв4r\/ڲN՘}-1-F$ݨ1zJ7j.#|Bp>X>` ~ʋ4˯o 6",z)<&rT}Sաl CzσJp;+NA] hJ?1/['uih_ ;㽴2 "LtAYa˚\hڶ|x|#4PՈwe,W%$(`CR9Z5?٫.3aBdjKeC˺ͫMJA_3<\ͧH> }Lomy< _@{9$Su,W;Q&t"ʥ@EA߸̇,rE啧^U6 T@A0=6 <亏AfME :Tip92/X, q*gWWAei+kF@qsbρ1A4A%^ѮW;N& F|$Lpc!ڣT@_>NA >NA'iDhz95 ߨw9K9kcJ;B6*DШ jЩcn4<0 E08!Q~n.vP:ћQ ClG,Z.r/|k4+FJh 7YRgIў%E{TM43XdQ:=ƞ +tl,U*BW8GsDEh.|{4fqHmǮ614Yu{o`"RJ˜R!ནzcR?#,[4uoDrfEBD`u\;MD%fڇST ]wz`|k,l) $u3o~FA$)%<XH~Gl,,lT7/[lLOk-7^zrsQܿ-1frr9'O?d1n\tM| 1c#d$ K3e.ջ!%ñ! Z?RD`Wi(G\pvݘ TXPW- 2 #9RHN9yXT('x< %O >:"::.ֳG$2jWg6y8sDeG 'ɬ*u5'Q)9ӈ2VE笗:%̋rʦ}L"G1\-بI(/9#"él ci㹖h)FFqaYDj;Q)ي "i ,=?3EX•(׈PpbX `I4)}4c.2xJ$mrS(DP^I[,'1%+((hג(j  —į0ae ,(^\vP(q~/ee' jlfgc}4L(%LǦh ղE4L!HPPz9: k :hX81jpG3,9Juf4ȩ aB.]7&L& ^'e&$ B<Ea`kyG(zYpD&xjT: Ԧ%yc\_*j}ޘv c1X B!F0ǝ2sj ZFQ9JHovl6u+)(d[r@\]1a+U!n+E(6)JM@"uH$V&X3ꁳ ^6:vzݞB# A?F* EtbE#c:TUo4X KYHuay E,I @Qy%& Lɠttґ>Zr0)z?1Zlp+Q/! Vޕe*:`\XN׏i:OυȄNb]~&'h5o׌p_h&2/gֽ\pP{:\mX_:};=^]s92hghzP^7Xf%N\=vSDGy qeSfZ Y0``LWUG&8KRk}iejOdz֏f&*g .+Ɖߍ6&~k 14NvLͼ2 zvCSmyqwoC)5Ѻ:WVށ*,T˂:Wŧ]Zy.h<ڸӇcV~F[_H&YnW5gֹSqOeĖ6)c)ta6{jvVe2͇ |;M~0K=yuW o vY*Ibg&{#狎XMg|6irҼ~̝zqcq>[f< +yK|9*WL5Iȕ2CgpCHsBo4*jRJpi)3#,A_^ʰY~w \Ӄg^ݺ,qZS*ų#/V6_z"od`:=[d늒xܛW7fv3ٛu7fт&ivlKϦYΦ0D_VK,cwD&g,o(rZzOюi\m cmGGD JE7(Zy2BG5Q.` E7]SHoLC0-hzX(i]aFp)1AyXpF%^z/@U\ {!D%9%'*ζV0XgQ0D"kE{]dA}D>6+gQA{//@^ eˤ,eI s(JUk$KӞ!1:d"}`.䰔fGxLBY#j]EO**bq,>+#!.+XqCd:[|tY/2(ܫ{D`"bnj]H|H/6e>Y1,=+&=b).\h +ɥ1z@-_8!iՆ[AR s!byӇe[ >c`Bw\OkEĊjX(,1^|wj@iF2 Sxnp:D$ :x[ ?<^Npk}Sqx=EW`X0ftdZePJoʅ#CDiN`eRQ"9!6)r1\Ypv~?`QkRSDDr ~*%/ z?&{Z vwE% HU (k] x jDQ!։Yەȕl]Њ4"wRYMt0 GSuUmH9cXNLU!N'8w{ /AO8ڌĐ miMJCRj%TbE|Hr /uU%/y~(g{yg9n1 ' #m Z*׆J$NQU܂vT9,Cjy,ښԥ%E#,vfΜe&g▙G!~Qs>yVE<dy`ffG]&Opiv0)Bpo wyCbuZn DtP3O.Y Olt} @4E|>y͝WeJj2Vs(YZZ&c_]ƛb0I!Ê_Z+;xҾgW.sW.zЕK:qΥ \*%I.;I4 c2YSv_iQc A]yq5i aՐ?Ў iƎCy֝׍,r(S~B?eX$x"ϢH>F!:yhgIy|yAH5Â),՝%q?;)/lImȟFuyw!G.ᘃnSt|}ȿZrWxucA=ҸVne8b ih8DZ=Niv;%ܷ pc;mr cjljG;6ͷ J%S~2ZԽywC#>FgGT!; 0үa1]^MmXyğ y-ѝ+ߵsy@+QĄux2>>Ԏ>ގ1]r]rhT'(U Q"=kExnֱN(ڋOdCc+p'nVwtpo'w֏n]?,]?ց}h1$[Z#uF.~@lүυ0J C{N)9[PFk2&f>)@DUImI LBp*-O gai$uR@Ne$ܤHR, ˤp sÚ88 -Xb8[)7j~kq68:dpӍ3ݠR;#ᕫynsI\R:qtVo~?O. ߛd[o7X ^kL#fq9N:O UsE^'laɩzurrv+/jXWj\)F^$4d[!D&0 TɅ|q٩slJ8+Ǫ^$;jUv+α"S˹rTûDVȩGIV:-t]$#7ٻ\H9X/y'Xxe Wv W]$+YtcqE2ds'a\^!jq1 Ggf@eK79YXts岋D5W\"?6득}CQإIW.fJ-Sx!nCɆʇ _Kf9i QET-鱍\l,f7גOHIEI!ChT{R8hu8-- F&\ttBk `W"}q9:ue:gyjR]ȤzoA%۰nq:VO(Lf/>O*(fb ('L29>y<sJLlHpܧ_>g,>Mps zRǟbG ޾rFqe2&PH^OWT. X]}FI 4)4O8 !Sѧ]^'O8 @ 'tsd߄;;uO^O1%M҇MsPZ]TL;Ň0އ"p>N“rS@!B( `UJ< Y_ *Z.BQRd\;Q$[{sŤű;P*{v>'ւW]tW;mkcow956V ,P ! W{aJ4Ei3zkv&k^vڍ\֫^X%^X^X[ky~qek5~Ԯ_ƋYnbFŌWk]_ ]_r__,.=cLP!?EAr%AXrԗPSax+%xE"=iV &F (;~ʫN9qdUԇ̜wzuo'3xFc:>!h "PXa WW*#kt/h˫N=k6l Uw%PI͡-xl9dia4/.¸N6ꕤw-a -rϜY /?m*LJCEuPZPG'6Z"*g ւ::9Xthj%RA>jp*Ԇrj~+ւ::e9 dP+6;5\.C)])]RL^x+>{isðѡuyp]H$֍ |6i&$X\u1inp,5f^ۡ_A9L(I2<;cnj!ōTX9_3r"%SJ|lj$a`biPGtbXǺ/ W72m@BN\DdJ-\nLCn4#:kc/a4,oMքFԦ5 Z7un4#:kcWߝZukBBN\DdJc ;Ԅ$9VԻxc;o PkL~ٶ_QMLv%"moe[٤&i}~}xTkPB=[mVÚ<-Jb*IMXУ dw m/:2Է} HU7 4 bә|1K֘sÚ%=‚5ܤ&jfhj̊j֘s:ⲭ15f5ܵ1+Ed[cnk̍j̊8F;'5擨1k}֘515F5r}|5f-(kkmQM֨Ƭ֘O 2iFR֘X#s[cnR0RjAs[cnTFGWc֘0֘O 2u5V5ܨ&(Ƭ bm17 I|t5fMRm17 ȣ1kf[cnk̍jDwYSԞcnkjEv[L)Um17 t}֘)15F5oZOg s[cnTWcfT֘ L.ךW8g7LPb/YE߆7K \e<ϩ9M9FK}.!GuS\Ȝ)l6.lB  /b 3E50Ő@TtP@JUzy`~G "nQK;`A۞Ï4<mg1|Jevբ7*E6hJoci*8Oܸ4"ueMGɰݖ?%-ߗViPSM(ͅu8!0Rs1+lB@86l8B8_ rFPRq0S(Q 5ڠH&T 3cLYHW@B"ZF > pFaIEH.rX< ǒP? -)l(Ld4a˜<cRK.=6޹D>wYBuH B~H)ˑsF9 << >8(I s)ne΄$+$$5IH*iZ3Khʵ#aA`b3!x4gUd嵠z< 88dK` 9̿!F ϳ:eJ94r)]g O^ƋċvB|f HH A08rsEv.gM6!k^D1$ 9#H8*{ D w Yf< G >Jԕnq>zE+/ =6%CqLMGlW91 7yx<0:"MlQw?̹,A/; #^¯~9ntvT'sv6uށ S'{[{s2 lsnPp0bw'wZ\,;ɧlbn=x&st0j6[$N~:/rr%dnDQ LQ{(ScP_.CyuWR^uJ_SqJ2-Ʒ 2?CIy WW*#7iىzQozyj^Ty:]f)P53Us0 ٨WB8>o :32DN\凰)y1}s"k7;=%&+/כNz2ۏ8|׽lnVx&?"d}JʲQvp̼DF=y[$w4 ۋ-G}L&i?i:g|aؗI?}]9YlEH`̩̽{5osR\p0g=|^47W`Ίanv2 0ι7 r 6Aϩ9spw{3w7|Nm80`1p H'G orTN$)K~@Y\I;^_GRI_kBi^:bd+ y.r!B5- GH OW`y$HzMfyy zgSמuXPc(NWKf켞>p^̨%oC<)Lw _ ,s-|4&He[7"/ˇrc.gRE\]OQڸ>UnPU:܏HRJ?zS)D ^NF /TXaiƆ2 fj̐[Ya2aB$9P#@j6)6bh)N$VmF ~@x=?L*'+ٝMjvDP~XAր^r:3qy<1t VySpT g*CZLYt<FqsEH+r DSxhٵ0Ur051]aEqk`%eMsn,2k#XjPTHbp3/ER-aC4-)X2ZiYXؼ}S-c8rfͱ#o П*>@X Ħ5Qԕ J-fV-js&F0Yo=S80 3@Ěg)*D2@ш?o!C_4iac5)ͺ 龄M)@ Yш+ɤFb#e Y%Peb9ܻH[4͗fN.bȮvv25 ~k?~Ѫi~cr )[#֏i$AAwrr*ieb3G=yD.+.Bc MP f:qlw#gf,sjNӚ.~M~ ^M7g|ngS2Iʘ-1 {,Ϟz,|* sT1L/wίr+a/h#(FlPNƅ }uo-|uTѨoy7*_WO9{/1=r;MF l $c1u2?QWLQn9ੇX2t|c8Y=l('ɺ&Ɗ`MXT¬)"tw~H֚[s ©zwc#AR>.|&"iܑL4&<.w+ l@Jڕ ;-2|̼W9'F!7xZ,}f-\rq.[\kI3&fsJJwM}s J('SaxAKTF%yŎw4'[1%u޾)Lo!O a->t S_>_}7D7Wp^U$OqLvɚw@ $O^IO>yc< ȫkLR'>#Սo!MzVT=Z:Z Ȅ@L2! iJ"$BL` 1fyPe$}e9Gu|Q5_\뭺& /; ̞s+gٔ4Ojp;FuEYBy/0aVsìU5OG#!oHU .Rt dN fHLK+ 彎_|~.yrN<[a$L5}ykmMLzZ?f;{Xyiz⭕Tl)'{]i1uFSL(ThTIYuQ:4GA"И*inH%㓎B:PΫfV{ΒncrK'vȵtn5_ͧe|+Q(ӽhޢ@%r :@O Q={n})IӼ-(Wث \LUO/'x7j.A8ej_i2lټ@Juw@>B yFR4ꏹvۘ g7Q)Wn~U6fPu bqeeG@EH b#HDv*k/T8+Tg.WL]=%3UcCv%Yqe:W׀A PHBf08H90I5B* Gӂ(9K_-2"qQ kINMWM&lXRrp6K\Pґ"xu=)~+|S ?7BړKq=0B%wMcnvs7g?\+..1adǽ?QvD?K@emnZ.::|h X;?/c9UTY_lFX:ZaДatlAjH*܃[G^S!g`/DM^ԛ ;ZEkjpp#E&Gj^PL67zo7٠%\>g+-IݺsyT0"% )9 _Av!5k"";ԩ=iVC,)#.3cqB:t8a0RI,W`ŖǚuceLUTTrA^%BCrFJY_gw RSyb͎w ^uv2T1Y@)yYZk`GeЮVUө/T:ҔvդT~ڛ\j,ZVXG5~S!f:ցV[MW2Zyr F'rOVPug&ykk_ME~5.*vI\@++-Yۃ+EO%uRP%”.'"|J 츟 6kOf %^S͆i#:CG1ּ;OvWAwӗ=mۏ1"{mlz Gs(\:oyvc-PMg7MgCGo~<< -$#/$sBH%m>M,wKmyԃ>,?*ho)ǫS g(o!|:GRj4r=6(+l3Q\[yǟp=A ={c9A%~7N[b/JPp4ɫ(Zz8.wJ۔hOlWtF8+Vt:yn{ݯ:?mkhۺ^x.^ROr<.PP(TŴSfd+@;^d(s LhzVW `|'Rt=f$װ Z 5 v~\6nve4RU T0ʤFT1 XL *%,8 ȥrFSdL7;tS:u6AiLMg}7vT5 H*#m_"C8"gp0kaֆê6\.pH;/KpT2Y%c T4%4A~vዚeقlⴅaluMlۛ W«QK&3aUu,]Hц(|GF4CI8KRyr"$z|`:nF˟/FkG7t9up,3tE:]&mȍ)QitJ.xr8 h(G`S>tf>50kj`fjbf.ǃd o1Iz*ܙ ^:AHk!M̀ mïTe6 0DL.뵒y$DZvrJIXh!4C0ooJ^tKٺ>#RY8H2hm<#%İs $r L)pSSo. :Mm(z G?2J'}toYxJQ.@iũ={n}_n!CW"uus߲xUOBl~ѧ5@1҆ih$GcN%htÙQd&*Okp(lJKl '2`JɇcceLJ*4 JHBd ܠb u,fzf CLE̴L33F)T[DEyf{mc 6M15Ō&6,l)W9;Ydٕ~"xu=)~sS F7rfƓlR.0B%=K] wػ.Ylvs~vFs]s>v< #;MoL$_r}-*S׋ KR Ũ녑ˆ`t1֔c.ZPmiu4WITF^S#'5ZBMU"C]uB_:PgTdcr: CEFTxPM#iˉ{޾=JqC2 ٪~?^Vٵ01%yiժ ?q<GQ T0_By0zhPzy“w/w!^&P^ݨ*{3Җ+זּ^淡t̷mZ)E=<](y\~ Hߌ6rD]2;i9LA~~_amLQg@@+րz1PSF%!V@riO@y 4)X69AR$^Z Ա)m*[@bT*4]7>Coz`lXB?,IǞ^F{ҡZx[8gق;?8XXGtͼy J9,I,WŤq璏$qQ<"{A*w7nSknW5E[rRG_:v;IIQ53$Zi43V{fHnEf >HTߞcܷj&ds&8\ kZr:.b~R38GY0h%eNGE­gD )k B{w(3EϴEcb9 W!Qr ZB /HRy+ 11Njs_Im(*h!\ !Q []KHl9J6i.D#Y\ |MJ*80\A';ޱb䣺rq3* H;#@Rs 9c)y{V "&9Psu58ݞ/QXn%J1S}𵩾aG@-f2 \ZZ#*0>$F!{r \ TH-9\ɉV;Xv3 7V'1o) v lfBP#-+=*]Bcޅ` jDŽGTBD}Ǫr{s Te"Nr&cړ I5HQGoC!4u%to: f9ߘrqL^|ze vU8vdjr@ox~WwLo@~ugmZ WCgpGp#q_GPʑ fz8ɻ~ˢ?컁daf# e]Q3n ǣXMQ><9qGj%Lǽ....<X61g*HF XJZK1;#lL efITNg:|oˇ8U׏+Ԙ&g%7.qj|x+ɉdiQN|p)d22YHڎՂ53=VJVKbۘ]TqcAgĠ- V{ժ~dA~md2i੿_b^Gp.xu!]+~J2& 裥|0Ggj֡&1glK &Hŀޓ1dlU `7"˽/vlyfLE 5\y ]0C[1{XnAͲ4_--4kRBt?.i 9j{cRFw-dge[ Cэ2rE xaU>nb2x%VgсF~T*׋KVn]j8Pu:qw'1""لA;Vy=:QD4wBj.2G|Th, ĆUֱxQEg4+i޸Lf 2;o5j$cTb,qM YL(6 ]j=T]ؙ<<nFX: )" sɨъ^謞/X@|-T3CmIQ̢o^Z!u~׸/?)c2e'=Sh|,keg,V4?8`pgD6gǡvTp|k_ g*1s-`vVWFuJA z Eok*~tHR?3(s̜]HmwvbPYXxX | M/ˏt߸,ctqO:>ې63Mg$ &ny9*~XBypYU=u?dή'kuF}]znU[z{CQht )U>=dkDO'2,) 3g~Vvo-H~p\t89";işk_j_RzxõT :G)o/KFW;ݨ GiٳA7D3V3г2bI(w}[Fʻf m{jbxߛa/!{:Xן Qd GVL`3i¤CCCA7(Z1O%=¤'I2m~Bf-od3Qs#W9TMe>7B+52_5/rɌJrG\eQ'Fz35J+)Ɓ;nfQc%qn2J}A",f) Bqr1aL\|wj*_-z0yiAf2R%GۦKCHԒ> A-gBp3HP /:*gTi"\J#1u3HV0+˄b8[[3HDv#ѠBnȥD,H5Ġʉ: 2Y9ns67Vk%E\gh؟ôԸuݏ˛O]&d 2ciSp5Bɢ֎9N 3َ݀wG\؈j^B\5~ ڑ26ѥP?pb„xm+1sfxh$ŠxC_ܢ"}fk,į2oL&AИXΡlƯ2䮓ܪdU,*RBE@1)vg0:E3-8s|> 8X._pPv:.6Fo̅03̈́Fqdbu<4TqFrKVh,y4p -+\n;(m2$N)#ͅ20RBURz86AC1Ą A R&JaR܇UhqeB9ʛnM+4 Hv%`9M{E>ii`:3D1S:ʙb&Eb nYو5k^By0"X5h*`7MA;ʹ |EG*wc 4ic=^^ E E E M5hW̚k/ lz 1X:dA`!Y:ixMg|P=VWN|C:*+[kANmX*$>h'B W1MFƓN Nb.e^ifh$|;awU{_jיt32:ՕΔ#+Ռ;,c^)-W9%4Q3<2BDl`3Z݊|yGG՝9)jEie%jH$%+s J63R:kGGrΠHY>#jX`4{%֙Ffsry8mQEZP+4L6:u]D>r+dtcd})>L{GTmo0B U##T}yy$m@B= lfش[;>Fiy$Dg'x;VEm\*;u>ޝbV>yH{GFziRsB܎-mg.-<8 Ni|c}C1ȋL},WQ:9(Q]'[Zmɑky!Y;jpoXy]%dzXNk_kVs ?^hhc&\_g]֭. e~mwq-hh[>Y).mEV[ը= jމը:4yVxcCMEqDecrpwQ[u};8L`f] N(l;Onk^8=Hx ]|bvC?*,Dvw5ٟpڳg*jوqeΞ~g/ň7K_䓻zJ@Iw3,h 5@5 讒XIJP#8i-RLn%U8m,rڙSpHtwLlL=uF=w&:zg6PPǎ  􈣀KG]<}iCUK^q ߾,u1 X0`'- O%iտۼp̛ :-t$0tK'm옭a;&.[k'Ҙffj@:z/$6np*ȴGC+9S"Gl;Z*-kӏR9Ԏ1ULڪQx#Vq{6ek0*cݭNpUfRG ௄3ߧhC[Dkbx +JJ[Ï,:J^ Zڣ ^ lq$B@ِ$#< qPOJZж}* ab|Pi.x 'ܜhE_UGõdQ·,E9hi s< 9d,F1!#g99 N`tI^v:fHĥX͋؞mG3/@;LRLDbII+mw'gY ɟNwҧ=`E\w (M[t)wħ`Vh!![0'-ög h%!_0#-C/"/pEYhpކb(.MtDٝKW1"C.R_L(W d!^'YU.>Gfdni40Z+<:@2 (Q[xi| &LS9؎5golVIQj LN7 E Fe,WRHH5h.ͧ\",K s9U?OFIIr"x~OX=ܠ)=HO)+U@JUH+,йJ 1 R`lF ap:v)?Ɂ:t|Cx,kDF '}IO1~]i 0P}?ezd:ԇ+26}'9U(z^<~w}GlZ*R͕L2M>~m8i|e8d  OzsxsCCmm.YҮ ҦTK*1nSˈ8ZgrHY=z_$v㙮mV6h/OUha5} eej-λ! q:hȡ, gZ3A 噆<Թ]h24馬?pyvah\"ɰ GJ$}Z7F -*破CҠL-]u,=:P/@މ̑,W9PezC+ =(Df1u&Ik D89 !*#C `)5;ݛ!=Ф"v=dՓL=d6C{ӈa` sHmiHSHeCDEukMij+k=dn Yl|5iSˣ?߱m|G@}r|$aN|&僮2)MNkLnPN0owsl7iH᤼w|Ƿ;t .=SHl"HHjM|'fkw5 Y&Yr#)p*y=N/}sŞ,עƥ~mT]URnԪX;/`RXWY?47>^1/-ErB*/ 2-s<Հ%匜USK`pdq|~,b,H ONzb/ [7߄^*uᯟõvnA2WM<8<"]hkI-5Sq:cW~yu?b~On-,AZ` n C[a-q}$xPlbtkpv{_g;"Vr|p8$>C g&K*FI䐾T$μgt~]Og00yIuLrg4^1{w<|؀ݢVVޣxL@~ߦw6˦ˁgcͶn,XaaW $2u .5 C\ꁖ/@8{Θ뢁 J*'Cg,1hW2G b *cΖ.^;{b]PLjy#7JRHJ85AY!6IE( 5BoRΡaE9 %eTVQ*2/fLǕL_1kX5 USM3d/ mʥ7{;è20L0)ä @aTA _ELESR(' ]0Lt [t!Ä4vVNxS٪v)D\?v  h_;R"Lb-!<9@$T~'](gL9J 3}wsm>\_nـɟzѾ{<3ɚfgtXnoz7O4EHm(?8)LX2묊_SiURڂw\.e=Ak\б,4Ou+,Fm{8lh`7lץT{/ܵ\sN#46zUhvI ZݜII{c-L/?]]^?ZAK U5}D}jda2c$z㻏տQcĝ~ז{WIׇ;LGw0a.Eey). dd b`(K}Q)E) }~kˉE|g\rR,Or@`Jkgd;OYX,.&FI## r5 !ږC *`oPH+*_m Mx@Є;ѻ_1vVgc 2^HF0mrNTh<ŒH-dYN>^VC[a<`z|ޙZ"$3K Gq6+1]_Й qRVoBp pѼٕ 6>\y殍 h^Dm[<,'e}x[" 9`MHv8Y}OQV_VvwЋYV_3pe97H@oKU;4ӲWVROVO XDF7K1YC5V6'37'j韝^~i G]2^ϗӵ5VgkU2D65%T-mHcjoz߮  QCL_~4>BFOh"c\ҷk>lf Z0D&&2E 4k Ib3"Ӗ9^LgE߬,W!@7Io15ſ V`oc*vO iwG7O\k 9B XÏcZF?&|9L|9L|9L|9e=XJbUTŢڨ#,3;<(])(FevqۉZE|gf֩3]7]9gş4 A Vx㌶h`B@S)Yw0fpv1Zm3󊐓Rfϲrݗogq\Ṷ6tU ba2Fo834,pZRsIJez>߭m(l /mUjpNNkknFS[u*?l%ޭM6x}Y0d&Ȑ/~JD 1I<Ė=in/~,dr4ApѽGc+m_IK]9YDF, 2L`P$Cّo<$*rJ#͉@)0Q{5ǝQH?tvZ$9mUXÞv u~ʃGQeC^kkx쫃?ayG;C^FQ/l+`1B$0a4rlc;ǶQ(&9:;aa@h-" XfR ZETǶ5U@[i ^~O&k$=sjw W PxuR ʹĎSBt,7&'j8JdHKj(-[]Ov&e{5k] m="RʫP#I. @qƵf! j?!@ ; M s((橧W5zV s%po(Ҧ :+`,E*mZLzY\ *Ys%01D+:Xnŏ6 -g[wqK;) =NӞRTmҳ (,q~21w}iZZ @pwD[z4Qlt!:xf.ܣ9^WpC^j[hiki;*d崘26t kY6\Oϕm6(G28B^xbvIqD8e}vXU+we>ڍKF7 /frEݚXK݅,W/Yqjcbͮ./|W :[IDE_}ӵpl׳ pC'^m'FQ~C:*{#.sf;lN9.|G܉PB݆J==R>Z 뭕Z)?Juy7,b4YQ)e҂Q,w]RAӷ:OuTv E#{',)Ap xk֦!kX^%/lͦoVI3 V*3g-'/5ĒT+88Bx~|?..¤‘P&zϗs|]Fe!? .U~>-߇ٹ~pҪ[sI\\~TD_?}pVo>엲q/qKKDQn̍5U,d6T,Ć(v<$$5k4*sЅG> z\(c>?T!%d3TJrS!D AQ$ XOTS$7 ͍OXJkdINiDZ( I3 mMLC?)_Hk1CbdƘ0W*b8'1UdG .Z~bR15&|qY$e1rپ&>Ưxۇi1 B0 5T؆Y(rBm-]hsS-$б4@Lmܘ;{ !,MV_6$7zj +>3 [Nmچi82`_!(F5-rL =v^' A(zLBhm"!8` nSu {' y#/'ҘR!7x=}0 Q5m%+0*bXQxF1^N'^3} R=!\su?~A2ꬄF8%Q@QvQY!ر<'8G:acF q5zX B (xD˜"|k>ZϪhҘLL϶fIJb ŅQTTSTNs"4X( + tпS~|Bo(ݝ_?Rty:KuRqX2,k6B}4%1<b fgk RyoGQJ޲P0lW0kF$RdW!S䲆1{IA`K@ţ"W+Bo% UvMnW3F1/XaKϥN:}@ ܩT݅]o&u~5ӵ"y֑c~Ia s]P)ĝܭ8bl^> v9Ɍ t?5 Vub!C痑 ;ݯ7>UY~Zͯ./J~cx%e_.FfU>m?Qd6w@3RzS`.J\g˛%s6wy7-oc+Mo.p 8\(sLT& qJgదA.|ZO7kd.l,*v#Q d[ Die2IQ~!%#0^냗vaxueoz65Hk~K@#&wnv 3tI-L8Cd!V:?[ͤu2sDˌ\e3VLc>U['ya\ϩ. =t ~Xt/"zLQ7XA8V#c֫_2 9ja 8TbT9,nje( +'5lyZuzVcٍalְ8}a]Q LljCB 8@pR\TN}xʭ<$'FH}F翾?O߿jthBomgPn|1 [ǹ0^NX祖FrٮY^:yH$~\9 R0)#J.z ; ^оһhn_Aw߆3`IQI Ԁ%tHc;ͻE3 ~Y|Lɝ3}s{acW;'i0@@v }sY?+:!Zbnx8E*7+R .F#hFu2 ˬ1Ts *;IUV2o4oi%t{@8vokgu? q #Ԥ:>qlvuy+[]17l3NqfC&HD #mpsNq@*} qISTN(jM=?G53ZSs|PlQ-Mi٧ )k]|?y f##U׸4nOG ϏG:1?|_6ZM8~sZiQ <Ũ$w-ןQpd3SEhY]*%M9ĿTBd  Z6ElNń %ȆCx\\ÜJF@e$XmQ$Ǵ71ޚ*6]SO"krtpC/󣽝}R7=R<߭#j\mYˉ(sG~N ˪D@\QV0$"#>pVU`Cۯh0ƖU 1:/ի5;ݓf"G gռ2n^MQ0p .o.< Oߌ1[%8cXX|v?5=Ht@л?oK?HF)yUxeG>[\Op%6!Z Էa_oINq"?JOKq7vUMDGHߵu5ew2ߏog VsO>;/ y y 1„p{OL5[LN aD;rߟc o#5ZkwB{EKcY% fjUl1P>8KjV72w' ~8㢅>j;iU09ce[u czTV } R2} [[#Cc&<8a/[XA*qӏ)IIf'_W.I7|j1 [Dnd6Hw8R4_o ^D(w w?miz~%E_ $> 9 ־!ٔ'q43Yş-T@fL1㺘4i# 6?vN [#ү҅v8*TdZ#\Jrwr'[RG>[! #x@Ia;3VDsD kiiԖR :"+E4*æitZި0[&ʽP$M"fpЎB;^XQB5&X%7J$e9K!bUҎX: -`}QPg\M'[M[F6֩…,?M&EeǘX=%/Z„T:yʹu.:zP#x,ʮDiNp6=u.jZg/݁Y`d']uS_eNHK(NS%jq|`P-Ž?{ɭ/̈"Y\dANI^| 4ojW4OHsJ6`J3WbXU79,-y9GnSƹ 3ɉZ22e}O4Ó{ .2"/2h= 2aÔ1p5KZxX|,r)9ԾIxikfʒ rδugߔafks67,A0xE'¥̔ ױUqz_pQ*VCG JpBK8uT>YƠ %1p H%$] ڭy֡.bQzF%߸ DZAGLo݊FP4V*Ģ"[Bq*Q0nz]X]ul 7mP﫺 κݒnB)1`@͙%hcr!ȰdEŜ@6'x&j{&xdZEogTǿ㯸u0XdU7|I>6d 91_)ڂ&%K6MIı9mv0B;%y6mj8¦"AM}"ByU3Eّ!"RNz.G;std]l NWZ<ߝmjW}ikj^!o|7t VE G1^!T8j^dռȪyU󢭚 _8FLL0FGZHt 7(D2MQ\h U]־|];i37z ސ|7{}lj;PǺQI9cF4Ƅmˤ`1r^CA|%:ՈרΩ%>fǬՆ77 quUC5ILh F@JT-'r sFE=FhVfEbuAX:l aSMoF_GrPhp+ ZH5 坮QmaAڨ0?9u$oRgL[# )VahV# ْXDALs%BÅI\(㢍2Yo%C7y%>H5Fj@-n9n:^!݊ޭx¨Q)X)[ +-dڢ#\Kg!j ThN0ujWPo6.Qvb e+<[=p K ;u<f@>OyHeUb݋j  " (U'*o j{-lWvLY6+&ۯ4ZH}2E{OI^%j#`u7n: vrV eQ]3ۻ}-ڗ1.Cxv2 4J"k}vZ${{#kwq_5>wFzvUJd>xw-7rƕKQ2)z[:021^pnZ%#4'hMfE(mZr5?p3Z+`; sn+rK2+PӊS4Nhs#pI/6'%mnfv35S,G(̄n^Ƅm,>r@f@Q*ڶ F{ ֡rkh{؅_H*ξ\?3Yg/j, g+-'':_6o1) p%5h'oS "wRqqR+ԝ D+]=|Gh.{s:iL=AtA=Zō,w1"aZl095Bq.SG8Sf٣-RT$W'FNݼ;v3|M+ay4|D﮴b8?!^/3P)M yەPbSam fBZD=vוf7 3O.)aDꉭa,7?77pin@F_Y_z}Q\Хܹ'N:C0#-;O]0؏q}/Qܶx:_]cv|3jdhm V&hvk~L*xH!XvZup_׳gx*,Z3Yʵ߽߼ ?- \'T:D.tzG14㹗bn0v% 'Zh*scDDIC^jeNml֭թ}GvD 3h֭ʠк!\E+tJVǩz[MH~֭թ}GvSZwΡ[=GauCC^SVj)XvNgu*6Y{oήoo>CG廣Q7Ӈw6W|R( ̟/߽9y(>}ICf ]MstWc 7fYߢt}ys)xwsџvLDfN 2NUf܇jlzMqM$}n/>E-U#SE+dr~[c n-q$jHo3 y߄6PjYho(`( I_SSSfpUֻ|>E t2"+EVmEo.w>5Z47Iyi@c.r䐄a)T`wф.}QUڗJzZ|*5&wU)SL ;KYn#R  S[WoLR(NBWp!$?ǣX<Li1pRM#0k"*xzǢ!# 6zt#P[gDT&J[H5^3j _jFYiZGOsh̒$r!ēKi X0A@t5cBQ V*?XoH ZCK͋N d)vMp֫`;^} M[$8k1(!=}G^m Y-yyB^>4䕫hNu{uCѲGnN;X#]RǩX=Rևr)'qn[7-NºbP:bݎy@=V=Rևr=XT49bOq*Jޟט3Wy-ƳûOMn+̔ѿ=mxu3\6Wg3s~~KmLwwr&:dޮY]a>fl>{`Gq'dOf/%4F?ԸĒt|rv>y9laH2M,LGi4.);@eHL̫hy{5^EGw4ZzKCLD֫y`"p(+yYRGfPBo0g۝%nN@0XrxvbBh d;5E-wmi!h)pdVppVlj˛ٳ2ݗ\8w| 8L2De/%'ۏVdZ_c[S+t૯u^EoC`g8G2i% bKDz,cMb>&%m%Wvpgd+{է-_yE#Rz&} N^br_}rVF/ !?4d֝K |[m5f,W?uBp^vY0U7rnKhsZ})Dj9L/Zk"ŋZlqFCJē7`3Q1a@ehS3ƙCg8+&}QqQɮ,Y_LADԌǤp2$0Q24G%dNhfksXEx>cYj7Ve횭bfՖD+6Vȩ*; ܠV@{&XcHQTNby{Zq '֕i5Ud~0nITp -5)r-YG̈́I.jAO99ح&đyhkC ym@ 3xduli"șFWe#d~lC\jݒ4V.=W{v(@{cKߋ(GB!iqgG&G4LR G? PiNOط3Ozhh-TϙsN獆K8|Ix/??a?s!o8},!j0dAX>u Nf~ Mm vHHS۳ˏ͵RA<]_f81L A-q|B" ĉD.,ՕTPk"'zqһ"IN $?simdR1ᬘ_%;VG[dq<.7ga9 YXnMŖMOE 󓘐NPX 芻[s@&Z //"|}ZZe`ȾV^]b\)׊+a'4R6l$){Փ-;6|H*OQDEKȁѥ+yHǼ#~# !N]Z5"H+(Kƀݼ;7քy8S 8Ŏ*xtݻz{eJjk4j_8 ΏvK)eX7 E>zx=]*mFT$yҴ۞t}G Lhe+;f*j Fg4Si&xjDSX$)9ˆ=5fN*Tk8qǃ駉9&?!O'g[FPt;ht2?nYfG5_Q =p$fp/ e%.+0gl%'Gl 3fނi][}$fkוrs#h{i{/HQ^Vx~򒼟bA E: 9p$33u|xk9:F>V2@iyhYnWNk1H4$IH[Ib@T~ԐЮL)C*p[8^p&Xi k'r\u6S t%agd"RP]"8]CD;fr@,,]Ǿd 8+:7跁B r#)drޫS`@ G,~s訾XU4 H'͟`~պW+ݮǕyفYԎۄttac 0)Fń0]|g^zzY6LpK`mV\ԕy'jGYjRXDgm(+k3wi$DR:@z^ bҩr xKJEX9KT]Y$T ɴ#ޘFPCP2oh%ud29 |Ĵ nKMcRJSZ 3LҢ3#Cg{d@ڃ6RЯ1\3 @/̸L X[!2?V$A=~v&%IAOH\ DkNĸq/; Z4IAN'}ccSń2?m;?~&Beo&(}l;ƉmQ*bk)q2e1Δ5H[LCS>%ZNmGj}bh=8(;QN4Bv=x1Ǩlmߦ?q~Z@TxL~7iiڣ# {Xҩݨj}QfK٧4ɷ°ެeoԺ $3.G9hx#yM.)PI!*[?rQ %? L Kv5*'$20*IO1-hEӧңqTl =\<ʀIJH2n4HYj"kq{H]sBWµe\^!T -8)MQR K˚SQxF ̠I "7A`QNzx eScfF+HIך\W͐$7eQؼJTI)cnIq2yzD(=Y٠>a:)ҠRFIe{ )Q lt6w41[cH"UϜGo5!ݞC9UG5%CdnȼV%Ƥ˲6eYu-F\K-.:j Ɖ8cdɕl#&ɔ\зcJ,)PJ֨Uͥ*"`,?5fP94OylW`SJ%`w."ѠN &]P sh9W N%rx6"jF=<-;IX><]%s  )R9Se]:ZP?y-`l듗P=\x$Ȝ *4O ӌZۆcɇ/'r>ȪUvJC'4tPJJ؍{)aSy}vl궵R{{^\hSUo*mT!D['1XV*aU:J*YK*~#IwVij1g+_g_\0.Yq'q=?ϛ :酪}{z_st1Ybn!a> ot( 9 "rxo< Ү S'p煞Å3]gp)E{bW1V)H%RtwqvL9,3BAd5lC#-PReE-5έgٖBܺ&{Z`ЕWP;[9.+] ˚Q)'+%:ۂ$"{2Sj+"nٻ Ԕᆁ1'PH}~am (j}Zg=Q*9Ql+{q]l[W4,CF|F}xDyI.56NwsLj[CoQ'h_f66s&GwŌ`_ٌeeጸUY8#׿T:I}3tU?by=?RC{/t\]lizn]N-%"*#8P)UtwɥV÷>I 6 I cI44SNU>Tw˥^dki.#ՓtpTtdD#Kٱ}% LeTr>Ȓ*rQ^ O+&p*b>HO|M!Q=gw@rt|;lO5Jxsf.'Z;b:+I+NB~$9R9KcxF`)$uZd lQ)=0<-Z/}l Gʥd;Ȓ2z{_ 8-?}|s:}l*r&GP7oϔD 76wPatQ+>̮oo6ceCWZU]:fR(YpE ?<3ҝ)Ǔ_yy|iiN^rA1(96Ȓ؞XFݛvsV;☹|?euyySyV&=~^UѪA0d nV=l5;AEADVRh(5(Е(>)E1UU+GmZ0 ]2_zWOQ#7v$on_>Wob51!RpS1W4e(2PRp~k@D2ckP5@ r|^:GqqO::'YcHwIXH2TM#/?4c&{/*vwT@[?b\:{o!x,O pq} n0?\FW$w.|ٓe?nȡG1_~<]Ko9+B_ ,Kd02mb3ٞWjɒ,`UN$2^%W1#FP{OMGrg_+.s\Έ}]iT˶6P/TDjO}3hRAܠjg'`]ڣJ "@Ҹ07&䁔3~s}P>' 7"3H"V?I"vCo빉8xh;$-G+JpUxv]Y.jM gW~YREKMI.)߯RM7-=h-JGbPO/ISf7^6{t7Sv6F</g\p4hVmf$q]}IT"jJC(k{VBsF[ݧebGU^Ip6<%7L: ^^'B-D!o)Imt3cMP|$ Gě} C{n` sk0R+}ж-+&LV6&!&"8PT-(N[My;` S#4ɀ.? +^U\Eo(5MF8%HG8TVKi1IdR.ҟ0G. , h/^?EGTx^Z Xu;=Ɩ60s'RvvxH"5d11Dq̣L9FH2.-keۛCPe *:GcQPbR)9'F5y"(ɒ&&(ol 5= G!@ubBX(C֦x3DnK*J2MRD)t6zrU$#nj3E| p421w9nՍVfhS&Z|SEi='7+' v~g6~seC{͚LU f.璥ςFИ(]Dכ0ՖbD$SB4 rr,?"0X КjaF s0Ў *>.\% n"PVגC!-sҿE#5]`5) Ig&.)W˺!lѹd 6P$%)&pa~O5Fh-:r?E a76(-3 ڠl_= ̫Uja(" ̓q4ZHanZIVpKSAJod4Hp冣e_nQ$$lC rCTOmnQ mw(z0к MƺWW)Ѐ7mP܊^lBHhMC"rU1b%$yylJ FSh\`35*F'`dhbNZ`嘹d_XZa}U`nEy\2O w8hH:hZu Zht~_Hv8jnm[y[Ud*YNѱQ9 IfA5sh tK c?J+CT @Lb{AׂgGq>gwd3WI"v7O ۘ@$HvuupjD:p I0H`.̴Svf(a+m%w8myiO*,&Ѿ~hyk=_ҦuSץc2~7/a6tv)Gg6$h]T>3Uw`nN!Jw[d#tIQ:SA6Pr. ;ql*!jid9QDNˆ< *euv2}]Aip|fe=Ƹ~uՔ"!#09F13xd^g4N*o)=dBڿniPSș(a%:t;ox-m ѮW]a6T9 0%4H<i76Zf8~jG؁|m);OULWH u@N 73r9et9IY/`OA{3xS1FbsFZE5Oxw "z[8U gM찑SDG!.k3cS%gy+؆IkLC:"VΛFA2M}s]eKۦWmo**#pzKd-@+ {ދT+5-]yk,o۞;(?}BzZnQlRb rhh @*li2?d!SXCcaakU5'#:h څaU:+2ژ^(66iؽb/ =3;G| UZ@r&o?^Zl!ϳ|xnr\ȖV\ZjGONzj o'<_~Oʌ.!Fɭ֥$ A%*g{ջ--Xʋ;=W'Wi#ܖW:St?g+}<lr퇅^ 6m}} kz+9AfV>NWr.[Opl5%(EatFhQ'gfo4聡mw:}=)g;3sB9->T*‘j_Hu/ځ԰tzbUya)7A}:i߄_ZiLR7о(WJǓeIɘ0z9`9/&.JB!Oa.ex|y=7?nN?|0Eg,(/ +|7@wq $S7ʦlY{X~+yF^q괍ߜ\|\~{v, lw8G/:AԬp|*Cf zD\&,E>QL@:rh ,LY(-}tO\u@]bo4W1*+Wޤ[#ZA(x@GNv:' 5:Q~FD|On p58\?OWF)§ j[: +Y(jX^4ig&Ơ*7$[g'$Hi R=](V|xǫJ}R{j:@mS8Ro]55)ZO18PNH3=eeѡ+0dqcuh#ȈVG2cc:hװ_nNm4eY@gFGRCOpkuI`jDhZ1Z:Q4 j<\Z K,`ƒ6ftR'8`K1ٟgF閑5zM%fy:.mbzǫtBqj`+9^R+k%QW߽VS^Bkj,wb!7&m2nZ_X穖%UR!`7iI.Y*_]֥A IԨ$=˼'Qʼ:OqNU2Xr( whܮ.;_ur~w:uPc8NZiNz .mA C%/|{S-5pˆ@nǩ^-L6V~9Hp{oޯRmۋd\}Fy&nkIcʤ!4䕫h/z=u\:ŠFu>u]E^oѺOϬnCh+W^:Gۃm;n}1Qbvw ֭#J6r])WSnXci߯Rm+:䂮nZ* 7pAWS]5+-1/`B-*ÑgKV~wwz]g_M:#7>}ӫ R^$;d7;]+je%ɒZ$D2 Y0J1rQJ5+lꝘw`s4KΒt0&dN7r"oع60 PB^"OQ~j+WPBן5cܞ򕁶z#r7O#kD9%MR9<լS:/9|X>aШ/_..ͦԬibd΢F6*C8cl$DHУqgEJA9kB{6ܖĒ w,$7䐭<]~7(yoNZL|g޳mkWv$ދMe ɴ4]왑m١,ٖck,S cDYD=gl@\AdP"EqgQ |gdSdnI&J3+(#DE` %LhԔ+jE)1I݋O&nI(ʽ;Z +L) 7o4:t7vEao_7E$/U7qQ۰0,y:W{=@ks]h\Zh15Lt<s}#匱P@FFk2x G6eh]tõ/3pȼ?8%*2@!xB[cbWJ$n {#p \4s.Ly+Pzn vsy[TJ$0">31rkAePDn0\@D p[J)1ӧ[ ] l6N:NOD+3UIgSowKbC&L9=(gFy_.&Ï~ʂ̀f >~2dLbk${Տp(<\ɣlY zi==\ilp;#O77VD~=8SmDz ԂOvQ-Sl.}1\(42̎Q%#b RA?94WY.NBU@?Iu?Q]5`ȄV(I8U;7t7#zR AƟ)5 7ǁ*8\fr4,d˦el\~,ф>\e\Ȑ:uA1RAr>Ô=m1ŎӘ6zQL2N @sօwL9Wg5^ gQ37|sbhYͮ1Р#ۉP9uH"4Ҙ{^x$Piؼ&&a<}燻y3&Z)]a8wW_=z|g1_/l-IyeՅ^ɹwz0q*r jx6fy8 n<;Ԫc/}x0%qiMѭAS `VvؽVBJBsekW;W5$A_Yh)hqN|C Ǚf}M@l4U[gn55ȾV$c\ t{ʈ͡.9_Zчmi2;̠ߟ}<5AJA߅)CU-v#L>h4Hh}IR-޶vk_ܕݺ#62%*}_=evkKi&mQzwڭ}!yvBB\Dd*hBAU>7&cv hH :뽐ċd\)k!>,/Z@m$BZs*MƀG X9G)CsÇ D†:FP N*I(FJ!(FK!G L S +$c[@-t3~ie\(s@!ͩ)iA% c)PDِT'Yܵ{#+1(}*@f}^ rhyvvF` AdY>Y#ޑ&׻!К:!B cĉC x@O: ! B'llH[Z(-.{eDb18pI瀞*+@"6pjcW*{,-(&`KRЃW>_TC{|Rʇ 1<=-D=t/lї7|||eV,9Cu* E&H=_')=h)s+\sr8hEb)-$x@VǓTi/ P,UƚHaIvtgZ/?QI'+r՚ aek7Z;?\hn~l)izaA}~DOhYôk h9H9)&9 t j+FOP>u`)4&wu`2 ˈSfY3z!5Pa_ Ă&BK$=&۵MIxA[">[3[) )kT(=0E)ɫq0{h9u/1 Եm}>|z=͚'=<!8ʙ{b%;k9/k`}3FJFm~OIdѺ![$'ɯRU^w|>y:$"B:a2NkVl e%{Z/Bd/RAJ˿hҕrypiͮMJم)iZuvb,|ڭ- RD;h}YE37Qu!!G. 2e 빛ED""JՏ[uP/BUe[C& !z!>_Z ~ʅ>d)ݩseƾ1=P/AIJ[J KRR1RҤt 6'R-gْ~ҴlԚj\JIRf0tdRCJA}5RhEYABJE/6IJXJEboDQ%+R\o̽i/:q=퍽fsO E*1 /QP's ]$ &,:u@=7?~7½LjDz6 m88!*n3NTr]'3xQB]p}R'ǮTf ڱD>U((|ߎV3SJ{GTuUklrUdy?H,ӜюESQF]  ࢯ*>Q:_~dþCJ6\ Lge"XƴZҥҥ2E+;x*t̚տ)tw ZѸ.$M+q̒T}r=e?ƞ8ӈ3ڙ[Μ,Mȷʘ7ɬ4熬=e "~=d+1Oߔ 5sd[pDQt@m{*6rQ;mrc9/r߇gO[*i,R9nmš &R SX 0eR~͈b+=ͬ_ j**)ZvFK] )3BM&rdަa2JS4L^tL S&TFǂ41g7;6sv@"UhcW`8W)przm ;0Ftyx0k5gD4pϼXhqjeF->e^ac&j˫{4QZ<C(Qk:v]s-nhE?#y8SC) Mͪ 8%N4/$y2V`56γ. ABNІtWa)=[ O3lpu6?=X.WzT)^p"_eqB,~eR. Pj V{s1Jsw I$񹄅\DY6qz|#x7.Ans9Sk{74w(˦>{qݘ#sR r7|l&`!7эMEM}qBW CkpU߆`aZ Duvar/home/core/zuul-output/logs/kubelet.log0000644000000000000000003530417115145355236017712 0ustar rootrootFeb 18 14:54:59 crc systemd[1]: Starting Kubernetes Kubelet... Feb 18 14:54:59 crc restorecon[4691]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Feb 18 14:54:59 crc restorecon[4691]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 18 14:54:59 crc restorecon[4691]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 18 14:54:59 crc restorecon[4691]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 18 14:54:59 crc restorecon[4691]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 18 14:54:59 crc restorecon[4691]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 18 14:54:59 crc restorecon[4691]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 18 14:54:59 crc restorecon[4691]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 18 14:54:59 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 18 14:54:59 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 18 14:54:59 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 18 14:54:59 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 18 14:54:59 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 18 14:54:59 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 18 14:54:59 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 18 14:54:59 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 18 14:54:59 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 18 14:54:59 crc restorecon[4691]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 18 14:54:59 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 18 14:54:59 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 18 14:54:59 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 18 14:54:59 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 18 14:54:59 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 18 14:55:00 crc restorecon[4691]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 18 14:55:00 crc restorecon[4691]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Feb 18 14:55:01 crc kubenswrapper[4896]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 18 14:55:01 crc kubenswrapper[4896]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Feb 18 14:55:01 crc kubenswrapper[4896]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 18 14:55:01 crc kubenswrapper[4896]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 18 14:55:01 crc kubenswrapper[4896]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Feb 18 14:55:01 crc kubenswrapper[4896]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.457393 4896 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.464858 4896 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.464889 4896 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.464900 4896 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.464911 4896 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.464922 4896 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.464933 4896 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.464941 4896 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.464952 4896 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.464961 4896 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.464969 4896 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.464976 4896 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.464984 4896 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.464992 4896 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465000 4896 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465007 4896 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465015 4896 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465023 4896 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465031 4896 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465039 4896 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465049 4896 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465078 4896 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465088 4896 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465097 4896 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465107 4896 feature_gate.go:330] unrecognized feature gate: Example Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465116 4896 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465126 4896 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465135 4896 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465144 4896 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465154 4896 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465164 4896 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465175 4896 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465185 4896 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465195 4896 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465205 4896 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465215 4896 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465224 4896 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465261 4896 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465271 4896 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465281 4896 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465289 4896 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465301 4896 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465310 4896 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465319 4896 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465327 4896 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465336 4896 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465348 4896 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465357 4896 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465367 4896 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465377 4896 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465388 4896 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465397 4896 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465443 4896 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465451 4896 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465458 4896 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465466 4896 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465476 4896 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465483 4896 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465491 4896 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465498 4896 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465506 4896 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465514 4896 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465521 4896 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465529 4896 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465537 4896 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465545 4896 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465553 4896 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465560 4896 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465570 4896 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465578 4896 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465586 4896 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.465594 4896 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.465800 4896 flags.go:64] FLAG: --address="0.0.0.0" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.465821 4896 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.465837 4896 flags.go:64] FLAG: --anonymous-auth="true" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.465848 4896 flags.go:64] FLAG: --application-metrics-count-limit="100" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.465860 4896 flags.go:64] FLAG: --authentication-token-webhook="false" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.465869 4896 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.465881 4896 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.465892 4896 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.465901 4896 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.465910 4896 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.465920 4896 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.465929 4896 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.465938 4896 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.465947 4896 flags.go:64] FLAG: --cgroup-root="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.465957 4896 flags.go:64] FLAG: --cgroups-per-qos="true" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.465965 4896 flags.go:64] FLAG: --client-ca-file="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.465974 4896 flags.go:64] FLAG: --cloud-config="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.465982 4896 flags.go:64] FLAG: --cloud-provider="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.465991 4896 flags.go:64] FLAG: --cluster-dns="[]" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466003 4896 flags.go:64] FLAG: --cluster-domain="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466013 4896 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466022 4896 flags.go:64] FLAG: --config-dir="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466031 4896 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466040 4896 flags.go:64] FLAG: --container-log-max-files="5" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466051 4896 flags.go:64] FLAG: --container-log-max-size="10Mi" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466060 4896 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466069 4896 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466078 4896 flags.go:64] FLAG: --containerd-namespace="k8s.io" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466088 4896 flags.go:64] FLAG: --contention-profiling="false" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466097 4896 flags.go:64] FLAG: --cpu-cfs-quota="true" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466106 4896 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466115 4896 flags.go:64] FLAG: --cpu-manager-policy="none" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466124 4896 flags.go:64] FLAG: --cpu-manager-policy-options="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466134 4896 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466144 4896 flags.go:64] FLAG: --enable-controller-attach-detach="true" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466153 4896 flags.go:64] FLAG: --enable-debugging-handlers="true" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466162 4896 flags.go:64] FLAG: --enable-load-reader="false" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466171 4896 flags.go:64] FLAG: --enable-server="true" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466180 4896 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466190 4896 flags.go:64] FLAG: --event-burst="100" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466200 4896 flags.go:64] FLAG: --event-qps="50" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466209 4896 flags.go:64] FLAG: --event-storage-age-limit="default=0" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466217 4896 flags.go:64] FLAG: --event-storage-event-limit="default=0" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466226 4896 flags.go:64] FLAG: --eviction-hard="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466263 4896 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466272 4896 flags.go:64] FLAG: --eviction-minimum-reclaim="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466281 4896 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466290 4896 flags.go:64] FLAG: --eviction-soft="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466299 4896 flags.go:64] FLAG: --eviction-soft-grace-period="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466308 4896 flags.go:64] FLAG: --exit-on-lock-contention="false" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466317 4896 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466326 4896 flags.go:64] FLAG: --experimental-mounter-path="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466334 4896 flags.go:64] FLAG: --fail-cgroupv1="false" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466343 4896 flags.go:64] FLAG: --fail-swap-on="true" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466352 4896 flags.go:64] FLAG: --feature-gates="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466363 4896 flags.go:64] FLAG: --file-check-frequency="20s" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466373 4896 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466383 4896 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466393 4896 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466402 4896 flags.go:64] FLAG: --healthz-port="10248" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466411 4896 flags.go:64] FLAG: --help="false" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466420 4896 flags.go:64] FLAG: --hostname-override="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466429 4896 flags.go:64] FLAG: --housekeeping-interval="10s" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466438 4896 flags.go:64] FLAG: --http-check-frequency="20s" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466447 4896 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466456 4896 flags.go:64] FLAG: --image-credential-provider-config="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466465 4896 flags.go:64] FLAG: --image-gc-high-threshold="85" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466473 4896 flags.go:64] FLAG: --image-gc-low-threshold="80" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466482 4896 flags.go:64] FLAG: --image-service-endpoint="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466491 4896 flags.go:64] FLAG: --kernel-memcg-notification="false" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466499 4896 flags.go:64] FLAG: --kube-api-burst="100" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466508 4896 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466517 4896 flags.go:64] FLAG: --kube-api-qps="50" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466526 4896 flags.go:64] FLAG: --kube-reserved="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466535 4896 flags.go:64] FLAG: --kube-reserved-cgroup="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466543 4896 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466552 4896 flags.go:64] FLAG: --kubelet-cgroups="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466561 4896 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466570 4896 flags.go:64] FLAG: --lock-file="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466579 4896 flags.go:64] FLAG: --log-cadvisor-usage="false" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466587 4896 flags.go:64] FLAG: --log-flush-frequency="5s" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466596 4896 flags.go:64] FLAG: --log-json-info-buffer-size="0" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466609 4896 flags.go:64] FLAG: --log-json-split-stream="false" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466618 4896 flags.go:64] FLAG: --log-text-info-buffer-size="0" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466627 4896 flags.go:64] FLAG: --log-text-split-stream="false" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466635 4896 flags.go:64] FLAG: --logging-format="text" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466644 4896 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466654 4896 flags.go:64] FLAG: --make-iptables-util-chains="true" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466662 4896 flags.go:64] FLAG: --manifest-url="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466671 4896 flags.go:64] FLAG: --manifest-url-header="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466682 4896 flags.go:64] FLAG: --max-housekeeping-interval="15s" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466691 4896 flags.go:64] FLAG: --max-open-files="1000000" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466707 4896 flags.go:64] FLAG: --max-pods="110" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466717 4896 flags.go:64] FLAG: --maximum-dead-containers="-1" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466727 4896 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466736 4896 flags.go:64] FLAG: --memory-manager-policy="None" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466745 4896 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466754 4896 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466764 4896 flags.go:64] FLAG: --node-ip="192.168.126.11" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466773 4896 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466792 4896 flags.go:64] FLAG: --node-status-max-images="50" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466801 4896 flags.go:64] FLAG: --node-status-update-frequency="10s" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466811 4896 flags.go:64] FLAG: --oom-score-adj="-999" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466820 4896 flags.go:64] FLAG: --pod-cidr="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466829 4896 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466841 4896 flags.go:64] FLAG: --pod-manifest-path="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466850 4896 flags.go:64] FLAG: --pod-max-pids="-1" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466859 4896 flags.go:64] FLAG: --pods-per-core="0" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466868 4896 flags.go:64] FLAG: --port="10250" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466877 4896 flags.go:64] FLAG: --protect-kernel-defaults="false" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466886 4896 flags.go:64] FLAG: --provider-id="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466894 4896 flags.go:64] FLAG: --qos-reserved="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466903 4896 flags.go:64] FLAG: --read-only-port="10255" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466913 4896 flags.go:64] FLAG: --register-node="true" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466922 4896 flags.go:64] FLAG: --register-schedulable="true" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466930 4896 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466944 4896 flags.go:64] FLAG: --registry-burst="10" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466953 4896 flags.go:64] FLAG: --registry-qps="5" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466963 4896 flags.go:64] FLAG: --reserved-cpus="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466971 4896 flags.go:64] FLAG: --reserved-memory="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466981 4896 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.466991 4896 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.467001 4896 flags.go:64] FLAG: --rotate-certificates="false" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.467010 4896 flags.go:64] FLAG: --rotate-server-certificates="false" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.467019 4896 flags.go:64] FLAG: --runonce="false" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.467028 4896 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.467037 4896 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.467046 4896 flags.go:64] FLAG: --seccomp-default="false" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.467055 4896 flags.go:64] FLAG: --serialize-image-pulls="true" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.467064 4896 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.467073 4896 flags.go:64] FLAG: --storage-driver-db="cadvisor" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.467082 4896 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.467091 4896 flags.go:64] FLAG: --storage-driver-password="root" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.467100 4896 flags.go:64] FLAG: --storage-driver-secure="false" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.467110 4896 flags.go:64] FLAG: --storage-driver-table="stats" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.467118 4896 flags.go:64] FLAG: --storage-driver-user="root" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.467127 4896 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.467136 4896 flags.go:64] FLAG: --sync-frequency="1m0s" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.467145 4896 flags.go:64] FLAG: --system-cgroups="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.467154 4896 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.467167 4896 flags.go:64] FLAG: --system-reserved-cgroup="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.467176 4896 flags.go:64] FLAG: --tls-cert-file="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.467185 4896 flags.go:64] FLAG: --tls-cipher-suites="[]" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.467196 4896 flags.go:64] FLAG: --tls-min-version="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.467206 4896 flags.go:64] FLAG: --tls-private-key-file="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.467215 4896 flags.go:64] FLAG: --topology-manager-policy="none" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.467224 4896 flags.go:64] FLAG: --topology-manager-policy-options="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.467254 4896 flags.go:64] FLAG: --topology-manager-scope="container" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.467264 4896 flags.go:64] FLAG: --v="2" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.467275 4896 flags.go:64] FLAG: --version="false" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.467287 4896 flags.go:64] FLAG: --vmodule="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.467297 4896 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.467307 4896 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467500 4896 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467509 4896 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467548 4896 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467560 4896 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467571 4896 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467579 4896 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467588 4896 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467597 4896 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467606 4896 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467614 4896 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467622 4896 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467631 4896 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467641 4896 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467649 4896 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467657 4896 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467666 4896 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467674 4896 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467682 4896 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467689 4896 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467700 4896 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467709 4896 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467718 4896 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467726 4896 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467736 4896 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467746 4896 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467754 4896 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467762 4896 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467776 4896 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467784 4896 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467792 4896 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467800 4896 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467807 4896 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467815 4896 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467823 4896 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467831 4896 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467839 4896 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467847 4896 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467854 4896 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467862 4896 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467870 4896 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467877 4896 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467887 4896 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467897 4896 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467906 4896 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467915 4896 feature_gate.go:330] unrecognized feature gate: Example Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467924 4896 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467933 4896 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467941 4896 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467949 4896 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467957 4896 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467965 4896 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467973 4896 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467980 4896 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467988 4896 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.467996 4896 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.468003 4896 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.468012 4896 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.468020 4896 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.468027 4896 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.468036 4896 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.468044 4896 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.468051 4896 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.468059 4896 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.468067 4896 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.468074 4896 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.468082 4896 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.468090 4896 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.468098 4896 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.468106 4896 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.468114 4896 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.468122 4896 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.472952 4896 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.490716 4896 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.491130 4896 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491329 4896 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491358 4896 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491370 4896 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491382 4896 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491392 4896 feature_gate.go:330] unrecognized feature gate: Example Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491402 4896 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491410 4896 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491418 4896 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491426 4896 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491434 4896 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491442 4896 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491452 4896 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491462 4896 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491472 4896 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491481 4896 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491491 4896 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491527 4896 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491571 4896 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491580 4896 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491588 4896 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491595 4896 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491603 4896 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491610 4896 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491618 4896 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491626 4896 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491637 4896 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491645 4896 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491653 4896 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491661 4896 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491669 4896 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491680 4896 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491691 4896 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491702 4896 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491714 4896 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491722 4896 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491731 4896 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491738 4896 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491746 4896 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491754 4896 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491762 4896 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491769 4896 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491777 4896 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491784 4896 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491792 4896 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491800 4896 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491807 4896 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491815 4896 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491823 4896 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491831 4896 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491838 4896 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491846 4896 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491857 4896 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491867 4896 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491876 4896 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491884 4896 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491894 4896 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491901 4896 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491909 4896 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491920 4896 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491930 4896 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491938 4896 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491947 4896 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491955 4896 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491963 4896 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491970 4896 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491978 4896 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491986 4896 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.491995 4896 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492002 4896 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492010 4896 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492017 4896 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.492031 4896 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492301 4896 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492317 4896 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492325 4896 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492335 4896 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492342 4896 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492350 4896 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492358 4896 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492366 4896 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492374 4896 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492382 4896 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492390 4896 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492397 4896 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492405 4896 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492413 4896 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492420 4896 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492431 4896 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492440 4896 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492449 4896 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492457 4896 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492465 4896 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492473 4896 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492481 4896 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492488 4896 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492496 4896 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492504 4896 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492511 4896 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492522 4896 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492533 4896 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492542 4896 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492551 4896 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492559 4896 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492568 4896 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492576 4896 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492585 4896 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492594 4896 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492602 4896 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492610 4896 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492618 4896 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492626 4896 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492634 4896 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492642 4896 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492650 4896 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492658 4896 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492665 4896 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492673 4896 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492681 4896 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492690 4896 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492697 4896 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492705 4896 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492713 4896 feature_gate.go:330] unrecognized feature gate: Example Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492721 4896 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492728 4896 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492736 4896 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492744 4896 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492751 4896 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492759 4896 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492771 4896 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492781 4896 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492790 4896 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492800 4896 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492809 4896 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492818 4896 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492826 4896 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492834 4896 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492842 4896 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492851 4896 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492860 4896 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492867 4896 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492875 4896 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492885 4896 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.492895 4896 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.492907 4896 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.495699 4896 server.go:940] "Client rotation is on, will bootstrap in background" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.504655 4896 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.504789 4896 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.514145 4896 server.go:997] "Starting client certificate rotation" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.514177 4896 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.514348 4896 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-10 09:23:26.709196596 +0000 UTC Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.514440 4896 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.594891 4896 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.606269 4896 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 18 14:55:01 crc kubenswrapper[4896]: E0218 14:55:01.608582 4896 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.156:6443: connect: connection refused" logger="UnhandledError" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.645351 4896 log.go:25] "Validated CRI v1 runtime API" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.765726 4896 log.go:25] "Validated CRI v1 image API" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.769500 4896 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.778773 4896 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-02-18-14-51-11-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.778863 4896 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.795878 4896 manager.go:217] Machine: {Timestamp:2026-02-18 14:55:01.792126213 +0000 UTC m=+0.721460446 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:d255e81f-4350-4697-bbdf-3f95dec57eda BootID:3d8f65ef-15b5-4cc3-be66-c8dd6cb956e9 Filesystems:[{Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:65:43:ed Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:65:43:ed Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:ac:4f:f2 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:39:ea:f1 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:e7:c9:f9 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:dc:89:6e Speed:-1 Mtu:1496} {Name:eth10 MacAddress:3a:fd:28:b6:68:e5 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:52:8b:69:01:3a:e1 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.796082 4896 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.796218 4896 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.799799 4896 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.799953 4896 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.799988 4896 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.800167 4896 topology_manager.go:138] "Creating topology manager with none policy" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.800175 4896 container_manager_linux.go:303] "Creating device plugin manager" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.802576 4896 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.802606 4896 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.802774 4896 state_mem.go:36] "Initialized new in-memory state store" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.802849 4896 server.go:1245] "Using root directory" path="/var/lib/kubelet" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.811919 4896 kubelet.go:418] "Attempting to sync node with API server" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.811943 4896 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.811963 4896 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.811975 4896 kubelet.go:324] "Adding apiserver pod source" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.812032 4896 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.820120 4896 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.821066 4896 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.825293 4896 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.156:6443: connect: connection refused Feb 18 14:55:01 crc kubenswrapper[4896]: E0218 14:55:01.825354 4896 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.156:6443: connect: connection refused" logger="UnhandledError" Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.825454 4896 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.156:6443: connect: connection refused Feb 18 14:55:01 crc kubenswrapper[4896]: E0218 14:55:01.825580 4896 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.156:6443: connect: connection refused" logger="UnhandledError" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.827781 4896 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.830510 4896 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.830533 4896 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.830540 4896 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.830547 4896 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.830590 4896 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.830598 4896 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.830605 4896 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.830615 4896 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.830625 4896 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.830632 4896 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.830651 4896 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.830658 4896 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.832513 4896 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.832971 4896 server.go:1280] "Started kubelet" Feb 18 14:55:01 crc systemd[1]: Started Kubernetes Kubelet. Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.836295 4896 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.836211 4896 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.840796 4896 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.842870 4896 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.156:6443: connect: connection refused Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.845178 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.845286 4896 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.845364 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 05:46:58.441933617 +0000 UTC Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.845526 4896 volume_manager.go:287] "The desired_state_of_world populator starts" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.845568 4896 volume_manager.go:289] "Starting Kubelet Volume Manager" Feb 18 14:55:01 crc kubenswrapper[4896]: E0218 14:55:01.845556 4896 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.845651 4896 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Feb 18 14:55:01 crc kubenswrapper[4896]: E0218 14:55:01.846496 4896 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.156:6443: connect: connection refused" interval="200ms" Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.846659 4896 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.156:6443: connect: connection refused Feb 18 14:55:01 crc kubenswrapper[4896]: E0218 14:55:01.846794 4896 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.156:6443: connect: connection refused" logger="UnhandledError" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.847071 4896 factory.go:55] Registering systemd factory Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.847119 4896 factory.go:221] Registration of the systemd container factory successfully Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.847883 4896 factory.go:153] Registering CRI-O factory Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.847951 4896 factory.go:221] Registration of the crio container factory successfully Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.848112 4896 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.848158 4896 factory.go:103] Registering Raw factory Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.848197 4896 manager.go:1196] Started watching for new ooms in manager Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.848588 4896 server.go:460] "Adding debug handlers to kubelet server" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.849568 4896 manager.go:319] Starting recovery of all containers Feb 18 14:55:01 crc kubenswrapper[4896]: E0218 14:55:01.860504 4896 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.156:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.18955f049b771913 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-18 14:55:01.832939795 +0000 UTC m=+0.762274018,LastTimestamp:2026-02-18 14:55:01.832939795 +0000 UTC m=+0.762274018,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.871935 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.871986 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872000 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872012 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872024 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872037 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872048 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872057 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872068 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872078 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872088 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872101 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872116 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872131 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872142 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872155 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872173 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872190 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872199 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872211 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872276 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872289 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872300 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872311 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872328 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872340 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872353 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872365 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872378 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872390 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872401 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872413 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872425 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872438 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872448 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872456 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872464 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872473 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872481 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872489 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872500 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872511 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872521 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872530 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872541 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872550 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872559 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872567 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872578 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872589 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872599 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872610 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872622 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872633 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872642 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872651 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872661 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872670 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872680 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872689 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872699 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872707 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872717 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872726 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872734 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872744 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872753 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872764 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872773 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872782 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872792 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872800 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872810 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872819 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872828 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872838 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872849 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872857 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872867 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872876 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872884 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872893 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872903 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872911 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872922 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872931 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872940 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872949 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872959 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872968 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872977 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872986 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.872995 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.873004 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.873014 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.873024 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.873034 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.873042 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.873051 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.873059 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.873068 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.873076 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.873086 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.873095 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.873108 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.873119 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.874671 4896 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.874697 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.874711 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.874723 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.874734 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.874744 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.874754 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.874764 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.874774 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.874783 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.874792 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.874803 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.874812 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.874821 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.874830 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.874839 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.874848 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.874857 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.874868 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.874877 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.874887 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.874896 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.874908 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.874917 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.874926 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.874935 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.874945 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.874954 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.874963 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.874972 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.874980 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.874990 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.874999 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875007 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875017 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875026 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875037 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875047 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875102 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875112 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875122 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875131 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875140 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875151 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875162 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875177 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875190 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875200 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875211 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875223 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875250 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875260 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875269 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875276 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875284 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875293 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875302 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875310 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875320 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875341 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875350 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875359 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875369 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875377 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875385 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875394 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875403 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875412 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875420 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875430 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875440 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875449 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875457 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875466 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875475 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875486 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875494 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875503 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875511 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875520 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875529 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875538 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875547 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875555 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875564 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875580 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875591 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875599 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875608 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875617 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875625 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875635 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875644 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875653 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875663 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875671 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875680 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875689 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875699 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875707 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875716 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875725 4896 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875765 4896 reconstruct.go:97] "Volume reconstruction finished" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.875772 4896 reconciler.go:26] "Reconciler: start to sync state" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.884641 4896 manager.go:324] Recovery completed Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.897998 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.899632 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.899670 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.899682 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.900771 4896 cpu_manager.go:225] "Starting CPU manager" policy="none" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.900916 4896 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.901049 4896 state_mem.go:36] "Initialized new in-memory state store" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.924466 4896 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.926647 4896 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.926686 4896 status_manager.go:217] "Starting to sync pod status with apiserver" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.926746 4896 kubelet.go:2335] "Starting kubelet main sync loop" Feb 18 14:55:01 crc kubenswrapper[4896]: E0218 14:55:01.927030 4896 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 18 14:55:01 crc kubenswrapper[4896]: W0218 14:55:01.927349 4896 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.156:6443: connect: connection refused Feb 18 14:55:01 crc kubenswrapper[4896]: E0218 14:55:01.927430 4896 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.156:6443: connect: connection refused" logger="UnhandledError" Feb 18 14:55:01 crc kubenswrapper[4896]: E0218 14:55:01.945746 4896 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.979881 4896 policy_none.go:49] "None policy: Start" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.981099 4896 memory_manager.go:170] "Starting memorymanager" policy="None" Feb 18 14:55:01 crc kubenswrapper[4896]: I0218 14:55:01.981408 4896 state_mem.go:35] "Initializing new in-memory state store" Feb 18 14:55:02 crc kubenswrapper[4896]: E0218 14:55:02.029247 4896 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.035177 4896 manager.go:334] "Starting Device Plugin manager" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.035470 4896 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.035495 4896 server.go:79] "Starting device plugin registration server" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.036024 4896 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.036048 4896 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.036230 4896 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.036383 4896 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.036396 4896 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 18 14:55:02 crc kubenswrapper[4896]: E0218 14:55:02.043694 4896 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 18 14:55:02 crc kubenswrapper[4896]: E0218 14:55:02.047737 4896 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.156:6443: connect: connection refused" interval="400ms" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.137030 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.138316 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.138350 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.138361 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.138384 4896 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 18 14:55:02 crc kubenswrapper[4896]: E0218 14:55:02.138859 4896 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.156:6443: connect: connection refused" node="crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.229594 4896 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.229737 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.231094 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.231136 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.231150 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.231294 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.231471 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.231513 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.232105 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.232135 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.232144 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.232295 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.232396 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.232426 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.232897 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.232929 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.232940 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.232973 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.233070 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.233083 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.233210 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.233248 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.233088 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.233666 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.233689 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.233698 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.233769 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.233800 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.233836 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.233848 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.233875 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.233637 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.233899 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.233913 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.233854 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.234267 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.234287 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.234294 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.234364 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.234385 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.234391 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.234409 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.234394 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.234900 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.234915 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.234925 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.280475 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.280628 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.280679 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.280724 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.280763 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.280836 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.280933 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.281074 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.281198 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.281405 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.281540 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.281617 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.281664 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.281741 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.281863 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.339613 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.340914 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.340949 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.340959 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.340977 4896 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 18 14:55:02 crc kubenswrapper[4896]: E0218 14:55:02.341420 4896 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.156:6443: connect: connection refused" node="crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.382899 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.382972 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.383087 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.383123 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.383161 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.383220 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.383314 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.383208 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.383264 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.383357 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.383390 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.383378 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.383474 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.383577 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.383704 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.383773 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.383809 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.383839 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.383866 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.383893 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.383893 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.383954 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.383977 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.383960 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.383955 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.383958 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.384061 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.384080 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.384111 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.384323 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: E0218 14:55:02.448984 4896 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.156:6443: connect: connection refused" interval="800ms" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.581921 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.582993 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.593351 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.613637 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.620778 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 18 14:55:02 crc kubenswrapper[4896]: W0218 14:55:02.636360 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-dc0ba76d4dbed14cdd185851c22a9c98b3c19c0a6138e2de6cdba0462dc2eb4d WatchSource:0}: Error finding container dc0ba76d4dbed14cdd185851c22a9c98b3c19c0a6138e2de6cdba0462dc2eb4d: Status 404 returned error can't find the container with id dc0ba76d4dbed14cdd185851c22a9c98b3c19c0a6138e2de6cdba0462dc2eb4d Feb 18 14:55:02 crc kubenswrapper[4896]: W0218 14:55:02.642871 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-fc69bd6d2393aa2928bc11f2668073c36a95e15bfe05b42860c5edb4645bca78 WatchSource:0}: Error finding container fc69bd6d2393aa2928bc11f2668073c36a95e15bfe05b42860c5edb4645bca78: Status 404 returned error can't find the container with id fc69bd6d2393aa2928bc11f2668073c36a95e15bfe05b42860c5edb4645bca78 Feb 18 14:55:02 crc kubenswrapper[4896]: W0218 14:55:02.644405 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-f192d2872819e84f43ee54f6824498719287c10d6ef13ab28ba75516719f0c10 WatchSource:0}: Error finding container f192d2872819e84f43ee54f6824498719287c10d6ef13ab28ba75516719f0c10: Status 404 returned error can't find the container with id f192d2872819e84f43ee54f6824498719287c10d6ef13ab28ba75516719f0c10 Feb 18 14:55:02 crc kubenswrapper[4896]: W0218 14:55:02.654473 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-272b6e4e2490d3f921a90adf5f79155a2e00e693b70fbd4ac90edc28dd2cce68 WatchSource:0}: Error finding container 272b6e4e2490d3f921a90adf5f79155a2e00e693b70fbd4ac90edc28dd2cce68: Status 404 returned error can't find the container with id 272b6e4e2490d3f921a90adf5f79155a2e00e693b70fbd4ac90edc28dd2cce68 Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.741927 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.743221 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.743291 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.743308 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.743340 4896 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 18 14:55:02 crc kubenswrapper[4896]: E0218 14:55:02.743799 4896 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.156:6443: connect: connection refused" node="crc" Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.844112 4896 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.156:6443: connect: connection refused Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.846209 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 20:20:02.763477444 +0000 UTC Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.932123 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f192d2872819e84f43ee54f6824498719287c10d6ef13ab28ba75516719f0c10"} Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.933144 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fc69bd6d2393aa2928bc11f2668073c36a95e15bfe05b42860c5edb4645bca78"} Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.934193 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"dc0ba76d4dbed14cdd185851c22a9c98b3c19c0a6138e2de6cdba0462dc2eb4d"} Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.935712 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"3ef7c566b49400cf4af98f97df2ff165a2af6d4ff48aa53de10b457020bd6abe"} Feb 18 14:55:02 crc kubenswrapper[4896]: I0218 14:55:02.936867 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"272b6e4e2490d3f921a90adf5f79155a2e00e693b70fbd4ac90edc28dd2cce68"} Feb 18 14:55:03 crc kubenswrapper[4896]: W0218 14:55:03.008112 4896 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.156:6443: connect: connection refused Feb 18 14:55:03 crc kubenswrapper[4896]: E0218 14:55:03.008190 4896 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.156:6443: connect: connection refused" logger="UnhandledError" Feb 18 14:55:03 crc kubenswrapper[4896]: E0218 14:55:03.054499 4896 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.156:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.18955f049b771913 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-18 14:55:01.832939795 +0000 UTC m=+0.762274018,LastTimestamp:2026-02-18 14:55:01.832939795 +0000 UTC m=+0.762274018,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 18 14:55:03 crc kubenswrapper[4896]: W0218 14:55:03.103586 4896 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.156:6443: connect: connection refused Feb 18 14:55:03 crc kubenswrapper[4896]: E0218 14:55:03.103654 4896 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.156:6443: connect: connection refused" logger="UnhandledError" Feb 18 14:55:03 crc kubenswrapper[4896]: E0218 14:55:03.251437 4896 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.156:6443: connect: connection refused" interval="1.6s" Feb 18 14:55:03 crc kubenswrapper[4896]: W0218 14:55:03.285836 4896 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.156:6443: connect: connection refused Feb 18 14:55:03 crc kubenswrapper[4896]: E0218 14:55:03.285921 4896 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.156:6443: connect: connection refused" logger="UnhandledError" Feb 18 14:55:03 crc kubenswrapper[4896]: W0218 14:55:03.339936 4896 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.156:6443: connect: connection refused Feb 18 14:55:03 crc kubenswrapper[4896]: E0218 14:55:03.340028 4896 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.156:6443: connect: connection refused" logger="UnhandledError" Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.544678 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.546053 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.546096 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.546109 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.546131 4896 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 18 14:55:03 crc kubenswrapper[4896]: E0218 14:55:03.546528 4896 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.156:6443: connect: connection refused" node="crc" Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.652671 4896 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 18 14:55:03 crc kubenswrapper[4896]: E0218 14:55:03.653700 4896 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.156:6443: connect: connection refused" logger="UnhandledError" Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.844584 4896 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.156:6443: connect: connection refused Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.846828 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 08:22:48.309743091 +0000 UTC Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.942349 4896 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="3bdc64cbe4646c598d98e96a7ff130d52867c29c552541b1c7adea6530a582e9" exitCode=0 Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.942418 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"3bdc64cbe4646c598d98e96a7ff130d52867c29c552541b1c7adea6530a582e9"} Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.942517 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.944192 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.944222 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.944247 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.945895 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f"} Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.945942 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322"} Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.945956 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8"} Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.947619 4896 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1" exitCode=0 Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.947684 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1"} Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.947832 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.948937 4896 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38" exitCode=0 Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.949021 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38"} Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.949181 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.949211 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.949223 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.949437 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.950307 4896 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="1d04996a8b2f120111c7e4e1e3486e2f7ddc1564ceb73a44ae8bc10564be88e3" exitCode=0 Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.950357 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"1d04996a8b2f120111c7e4e1e3486e2f7ddc1564ceb73a44ae8bc10564be88e3"} Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.950397 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.951143 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.951174 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.951187 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.951320 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.951352 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.951362 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.952498 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.953590 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.953614 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:03 crc kubenswrapper[4896]: I0218 14:55:03.953625 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:04 crc kubenswrapper[4896]: I0218 14:55:04.844347 4896 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.156:6443: connect: connection refused Feb 18 14:55:04 crc kubenswrapper[4896]: I0218 14:55:04.847129 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 02:10:29.73563882 +0000 UTC Feb 18 14:55:04 crc kubenswrapper[4896]: E0218 14:55:04.853772 4896 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.156:6443: connect: connection refused" interval="3.2s" Feb 18 14:55:04 crc kubenswrapper[4896]: I0218 14:55:04.955416 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"ee345b360b9fdc888389b754b031d19a1e67a2814bbe15402ea87f44055adf9f"} Feb 18 14:55:04 crc kubenswrapper[4896]: I0218 14:55:04.955448 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:04 crc kubenswrapper[4896]: I0218 14:55:04.957463 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:04 crc kubenswrapper[4896]: I0218 14:55:04.957502 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:04 crc kubenswrapper[4896]: I0218 14:55:04.957513 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:04 crc kubenswrapper[4896]: I0218 14:55:04.960212 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7d26b99304cc56b1e016ba622e79f535b6dfd04c197b698dae1720a1ee44afab"} Feb 18 14:55:04 crc kubenswrapper[4896]: I0218 14:55:04.960265 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"3ec861caaac932e4489277e16914fa252647b155492fec4edeb789ea62e11eac"} Feb 18 14:55:04 crc kubenswrapper[4896]: I0218 14:55:04.960273 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:04 crc kubenswrapper[4896]: I0218 14:55:04.960281 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"fd5e328de9a4e016452b309a56258165a94fdd43c0e26f7444a4a765f3cd11c7"} Feb 18 14:55:04 crc kubenswrapper[4896]: I0218 14:55:04.961118 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:04 crc kubenswrapper[4896]: I0218 14:55:04.961480 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:04 crc kubenswrapper[4896]: I0218 14:55:04.961503 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:04 crc kubenswrapper[4896]: I0218 14:55:04.963530 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f"} Feb 18 14:55:04 crc kubenswrapper[4896]: I0218 14:55:04.963660 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:04 crc kubenswrapper[4896]: I0218 14:55:04.965458 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:04 crc kubenswrapper[4896]: I0218 14:55:04.965487 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:04 crc kubenswrapper[4896]: I0218 14:55:04.965501 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:04 crc kubenswrapper[4896]: I0218 14:55:04.965864 4896 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0" exitCode=0 Feb 18 14:55:04 crc kubenswrapper[4896]: I0218 14:55:04.965960 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0"} Feb 18 14:55:04 crc kubenswrapper[4896]: I0218 14:55:04.966011 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:04 crc kubenswrapper[4896]: I0218 14:55:04.967069 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:04 crc kubenswrapper[4896]: I0218 14:55:04.967123 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:04 crc kubenswrapper[4896]: I0218 14:55:04.967142 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:04 crc kubenswrapper[4896]: I0218 14:55:04.971572 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f"} Feb 18 14:55:04 crc kubenswrapper[4896]: I0218 14:55:04.971663 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d"} Feb 18 14:55:04 crc kubenswrapper[4896]: I0218 14:55:04.971707 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21"} Feb 18 14:55:04 crc kubenswrapper[4896]: I0218 14:55:04.971843 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed"} Feb 18 14:55:05 crc kubenswrapper[4896]: I0218 14:55:05.146630 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:05 crc kubenswrapper[4896]: I0218 14:55:05.148025 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:05 crc kubenswrapper[4896]: I0218 14:55:05.148072 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:05 crc kubenswrapper[4896]: I0218 14:55:05.148084 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:05 crc kubenswrapper[4896]: I0218 14:55:05.148122 4896 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 18 14:55:05 crc kubenswrapper[4896]: E0218 14:55:05.149079 4896 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.156:6443: connect: connection refused" node="crc" Feb 18 14:55:05 crc kubenswrapper[4896]: W0218 14:55:05.178683 4896 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.156:6443: connect: connection refused Feb 18 14:55:05 crc kubenswrapper[4896]: E0218 14:55:05.178799 4896 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.156:6443: connect: connection refused" logger="UnhandledError" Feb 18 14:55:05 crc kubenswrapper[4896]: W0218 14:55:05.218510 4896 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.156:6443: connect: connection refused Feb 18 14:55:05 crc kubenswrapper[4896]: E0218 14:55:05.218605 4896 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.156:6443: connect: connection refused" logger="UnhandledError" Feb 18 14:55:05 crc kubenswrapper[4896]: W0218 14:55:05.511863 4896 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.156:6443: connect: connection refused Feb 18 14:55:05 crc kubenswrapper[4896]: E0218 14:55:05.511953 4896 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.156:6443: connect: connection refused" logger="UnhandledError" Feb 18 14:55:05 crc kubenswrapper[4896]: I0218 14:55:05.582449 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 14:55:05 crc kubenswrapper[4896]: I0218 14:55:05.843622 4896 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.156:6443: connect: connection refused Feb 18 14:55:05 crc kubenswrapper[4896]: I0218 14:55:05.848165 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 18:33:49.126422744 +0000 UTC Feb 18 14:55:05 crc kubenswrapper[4896]: I0218 14:55:05.976323 4896 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65" exitCode=0 Feb 18 14:55:05 crc kubenswrapper[4896]: I0218 14:55:05.976393 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65"} Feb 18 14:55:05 crc kubenswrapper[4896]: I0218 14:55:05.976433 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:05 crc kubenswrapper[4896]: I0218 14:55:05.977177 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:05 crc kubenswrapper[4896]: I0218 14:55:05.977205 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:05 crc kubenswrapper[4896]: I0218 14:55:05.977355 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:05 crc kubenswrapper[4896]: I0218 14:55:05.979467 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"71647042503d25307e9924a70b6a802a8d0d0ff4a43eff413dbfa517c0f4eb97"} Feb 18 14:55:05 crc kubenswrapper[4896]: I0218 14:55:05.979485 4896 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 18 14:55:05 crc kubenswrapper[4896]: I0218 14:55:05.979529 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:05 crc kubenswrapper[4896]: I0218 14:55:05.979565 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:05 crc kubenswrapper[4896]: I0218 14:55:05.979529 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:05 crc kubenswrapper[4896]: I0218 14:55:05.979630 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:05 crc kubenswrapper[4896]: I0218 14:55:05.980448 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:05 crc kubenswrapper[4896]: I0218 14:55:05.980476 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:05 crc kubenswrapper[4896]: I0218 14:55:05.980513 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:05 crc kubenswrapper[4896]: I0218 14:55:05.980571 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:05 crc kubenswrapper[4896]: I0218 14:55:05.980616 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:05 crc kubenswrapper[4896]: I0218 14:55:05.980628 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:05 crc kubenswrapper[4896]: I0218 14:55:05.980705 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:05 crc kubenswrapper[4896]: I0218 14:55:05.980733 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:05 crc kubenswrapper[4896]: I0218 14:55:05.980745 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:05 crc kubenswrapper[4896]: I0218 14:55:05.981701 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:05 crc kubenswrapper[4896]: I0218 14:55:05.981729 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:05 crc kubenswrapper[4896]: I0218 14:55:05.981739 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:06 crc kubenswrapper[4896]: I0218 14:55:06.837931 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:55:06 crc kubenswrapper[4896]: I0218 14:55:06.849275 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 00:20:12.139382298 +0000 UTC Feb 18 14:55:06 crc kubenswrapper[4896]: I0218 14:55:06.985007 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8"} Feb 18 14:55:06 crc kubenswrapper[4896]: I0218 14:55:06.985048 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23"} Feb 18 14:55:06 crc kubenswrapper[4896]: I0218 14:55:06.985053 4896 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 18 14:55:06 crc kubenswrapper[4896]: I0218 14:55:06.985062 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283"} Feb 18 14:55:06 crc kubenswrapper[4896]: I0218 14:55:06.985071 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0"} Feb 18 14:55:06 crc kubenswrapper[4896]: I0218 14:55:06.985091 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:06 crc kubenswrapper[4896]: I0218 14:55:06.985099 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:06 crc kubenswrapper[4896]: I0218 14:55:06.986062 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:06 crc kubenswrapper[4896]: I0218 14:55:06.986098 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:06 crc kubenswrapper[4896]: I0218 14:55:06.986108 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:06 crc kubenswrapper[4896]: I0218 14:55:06.986067 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:06 crc kubenswrapper[4896]: I0218 14:55:06.986161 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:06 crc kubenswrapper[4896]: I0218 14:55:06.986172 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:07 crc kubenswrapper[4896]: I0218 14:55:07.307364 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 18 14:55:07 crc kubenswrapper[4896]: I0218 14:55:07.307628 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:07 crc kubenswrapper[4896]: I0218 14:55:07.309293 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:07 crc kubenswrapper[4896]: I0218 14:55:07.309391 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:07 crc kubenswrapper[4896]: I0218 14:55:07.309417 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:07 crc kubenswrapper[4896]: I0218 14:55:07.685694 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:55:07 crc kubenswrapper[4896]: I0218 14:55:07.849855 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 20:20:23.925354034 +0000 UTC Feb 18 14:55:07 crc kubenswrapper[4896]: I0218 14:55:07.951073 4896 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 18 14:55:07 crc kubenswrapper[4896]: I0218 14:55:07.999272 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:07 crc kubenswrapper[4896]: I0218 14:55:07.999723 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183"} Feb 18 14:55:08 crc kubenswrapper[4896]: I0218 14:55:07.999983 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:08 crc kubenswrapper[4896]: I0218 14:55:08.000071 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:08 crc kubenswrapper[4896]: I0218 14:55:08.000088 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:08 crc kubenswrapper[4896]: I0218 14:55:08.000447 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:08 crc kubenswrapper[4896]: I0218 14:55:08.001219 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:08 crc kubenswrapper[4896]: I0218 14:55:08.001322 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:08 crc kubenswrapper[4896]: I0218 14:55:08.001342 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:08 crc kubenswrapper[4896]: I0218 14:55:08.349599 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:08 crc kubenswrapper[4896]: I0218 14:55:08.350822 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:08 crc kubenswrapper[4896]: I0218 14:55:08.350861 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:08 crc kubenswrapper[4896]: I0218 14:55:08.350872 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:08 crc kubenswrapper[4896]: I0218 14:55:08.350900 4896 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 18 14:55:08 crc kubenswrapper[4896]: I0218 14:55:08.415921 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:55:08 crc kubenswrapper[4896]: I0218 14:55:08.851058 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 07:00:28.444672261 +0000 UTC Feb 18 14:55:09 crc kubenswrapper[4896]: I0218 14:55:09.003538 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:09 crc kubenswrapper[4896]: I0218 14:55:09.003538 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:09 crc kubenswrapper[4896]: I0218 14:55:09.005464 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:09 crc kubenswrapper[4896]: I0218 14:55:09.005492 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:09 crc kubenswrapper[4896]: I0218 14:55:09.005526 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:09 crc kubenswrapper[4896]: I0218 14:55:09.005529 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:09 crc kubenswrapper[4896]: I0218 14:55:09.005550 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:09 crc kubenswrapper[4896]: I0218 14:55:09.005557 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:09 crc kubenswrapper[4896]: I0218 14:55:09.306699 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Feb 18 14:55:09 crc kubenswrapper[4896]: I0218 14:55:09.851230 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 12:22:50.997374561 +0000 UTC Feb 18 14:55:10 crc kubenswrapper[4896]: I0218 14:55:10.005879 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:10 crc kubenswrapper[4896]: I0218 14:55:10.005921 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:10 crc kubenswrapper[4896]: I0218 14:55:10.007532 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:10 crc kubenswrapper[4896]: I0218 14:55:10.007581 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:10 crc kubenswrapper[4896]: I0218 14:55:10.007604 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:10 crc kubenswrapper[4896]: I0218 14:55:10.007656 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:10 crc kubenswrapper[4896]: I0218 14:55:10.007686 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:10 crc kubenswrapper[4896]: I0218 14:55:10.007698 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:10 crc kubenswrapper[4896]: I0218 14:55:10.852279 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 04:53:01.603659657 +0000 UTC Feb 18 14:55:11 crc kubenswrapper[4896]: I0218 14:55:11.030579 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 14:55:11 crc kubenswrapper[4896]: I0218 14:55:11.030739 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:11 crc kubenswrapper[4896]: I0218 14:55:11.031720 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:11 crc kubenswrapper[4896]: I0218 14:55:11.031742 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:11 crc kubenswrapper[4896]: I0218 14:55:11.031750 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:11 crc kubenswrapper[4896]: I0218 14:55:11.852482 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 17:28:02.034725172 +0000 UTC Feb 18 14:55:12 crc kubenswrapper[4896]: E0218 14:55:12.043923 4896 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 18 14:55:12 crc kubenswrapper[4896]: I0218 14:55:12.324153 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Feb 18 14:55:12 crc kubenswrapper[4896]: I0218 14:55:12.324463 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:12 crc kubenswrapper[4896]: I0218 14:55:12.325898 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:12 crc kubenswrapper[4896]: I0218 14:55:12.325943 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:12 crc kubenswrapper[4896]: I0218 14:55:12.325958 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:12 crc kubenswrapper[4896]: I0218 14:55:12.852617 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 15:51:29.855093515 +0000 UTC Feb 18 14:55:13 crc kubenswrapper[4896]: I0218 14:55:13.853851 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 06:24:18.37802877 +0000 UTC Feb 18 14:55:14 crc kubenswrapper[4896]: I0218 14:55:14.031624 4896 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 18 14:55:14 crc kubenswrapper[4896]: I0218 14:55:14.031735 4896 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 18 14:55:14 crc kubenswrapper[4896]: I0218 14:55:14.237841 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 14:55:14 crc kubenswrapper[4896]: I0218 14:55:14.238059 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:14 crc kubenswrapper[4896]: I0218 14:55:14.239866 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:14 crc kubenswrapper[4896]: I0218 14:55:14.239942 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:14 crc kubenswrapper[4896]: I0218 14:55:14.239966 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:14 crc kubenswrapper[4896]: I0218 14:55:14.854700 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 11:15:12.978279143 +0000 UTC Feb 18 14:55:14 crc kubenswrapper[4896]: I0218 14:55:14.955081 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 14:55:14 crc kubenswrapper[4896]: I0218 14:55:14.963533 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 14:55:15 crc kubenswrapper[4896]: I0218 14:55:15.018121 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:15 crc kubenswrapper[4896]: I0218 14:55:15.019387 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:15 crc kubenswrapper[4896]: I0218 14:55:15.019419 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:15 crc kubenswrapper[4896]: I0218 14:55:15.019429 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:15 crc kubenswrapper[4896]: I0218 14:55:15.025519 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 14:55:15 crc kubenswrapper[4896]: I0218 14:55:15.855564 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 23:14:17.999795503 +0000 UTC Feb 18 14:55:15 crc kubenswrapper[4896]: W0218 14:55:15.890315 4896 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Feb 18 14:55:15 crc kubenswrapper[4896]: I0218 14:55:15.890496 4896 trace.go:236] Trace[1483919086]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (18-Feb-2026 14:55:05.888) (total time: 10002ms): Feb 18 14:55:15 crc kubenswrapper[4896]: Trace[1483919086]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (14:55:15.890) Feb 18 14:55:15 crc kubenswrapper[4896]: Trace[1483919086]: [10.002036115s] [10.002036115s] END Feb 18 14:55:15 crc kubenswrapper[4896]: E0218 14:55:15.890539 4896 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Feb 18 14:55:16 crc kubenswrapper[4896]: I0218 14:55:16.020507 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:16 crc kubenswrapper[4896]: I0218 14:55:16.021573 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:16 crc kubenswrapper[4896]: I0218 14:55:16.021607 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:16 crc kubenswrapper[4896]: I0218 14:55:16.021618 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:16 crc kubenswrapper[4896]: I0218 14:55:16.844687 4896 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Feb 18 14:55:16 crc kubenswrapper[4896]: I0218 14:55:16.856163 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 16:10:55.874205288 +0000 UTC Feb 18 14:55:17 crc kubenswrapper[4896]: I0218 14:55:17.025288 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 18 14:55:17 crc kubenswrapper[4896]: I0218 14:55:17.027310 4896 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="71647042503d25307e9924a70b6a802a8d0d0ff4a43eff413dbfa517c0f4eb97" exitCode=255 Feb 18 14:55:17 crc kubenswrapper[4896]: I0218 14:55:17.027357 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"71647042503d25307e9924a70b6a802a8d0d0ff4a43eff413dbfa517c0f4eb97"} Feb 18 14:55:17 crc kubenswrapper[4896]: I0218 14:55:17.027435 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:17 crc kubenswrapper[4896]: I0218 14:55:17.027517 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:17 crc kubenswrapper[4896]: I0218 14:55:17.028643 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:17 crc kubenswrapper[4896]: I0218 14:55:17.028679 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:17 crc kubenswrapper[4896]: I0218 14:55:17.028693 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:17 crc kubenswrapper[4896]: I0218 14:55:17.028688 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:17 crc kubenswrapper[4896]: I0218 14:55:17.028757 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:17 crc kubenswrapper[4896]: I0218 14:55:17.028790 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:17 crc kubenswrapper[4896]: I0218 14:55:17.029270 4896 scope.go:117] "RemoveContainer" containerID="71647042503d25307e9924a70b6a802a8d0d0ff4a43eff413dbfa517c0f4eb97" Feb 18 14:55:17 crc kubenswrapper[4896]: I0218 14:55:17.294784 4896 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 18 14:55:17 crc kubenswrapper[4896]: I0218 14:55:17.294885 4896 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 18 14:55:17 crc kubenswrapper[4896]: I0218 14:55:17.302475 4896 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 18 14:55:17 crc kubenswrapper[4896]: I0218 14:55:17.302586 4896 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 18 14:55:17 crc kubenswrapper[4896]: I0218 14:55:17.856494 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 06:10:54.323219608 +0000 UTC Feb 18 14:55:18 crc kubenswrapper[4896]: I0218 14:55:18.032768 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 18 14:55:18 crc kubenswrapper[4896]: I0218 14:55:18.034846 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23"} Feb 18 14:55:18 crc kubenswrapper[4896]: I0218 14:55:18.034986 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:18 crc kubenswrapper[4896]: I0218 14:55:18.035925 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:18 crc kubenswrapper[4896]: I0218 14:55:18.035960 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:18 crc kubenswrapper[4896]: I0218 14:55:18.035969 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:18 crc kubenswrapper[4896]: I0218 14:55:18.419968 4896 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]log ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]etcd ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]poststarthook/start-apiserver-admission-initializer ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]poststarthook/openshift.io-api-request-count-filter ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]poststarthook/openshift.io-startkubeinformers ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]poststarthook/generic-apiserver-start-informers ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]poststarthook/priority-and-fairness-config-consumer ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]poststarthook/priority-and-fairness-filter ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]poststarthook/storage-object-count-tracker-hook ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]poststarthook/start-apiextensions-informers ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]poststarthook/start-apiextensions-controllers ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]poststarthook/crd-informer-synced ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]poststarthook/start-system-namespaces-controller ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]poststarthook/start-cluster-authentication-info-controller ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]poststarthook/start-legacy-token-tracking-controller ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]poststarthook/start-service-ip-repair-controllers ok Feb 18 14:55:18 crc kubenswrapper[4896]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Feb 18 14:55:18 crc kubenswrapper[4896]: [+]poststarthook/scheduling/bootstrap-system-priority-classes ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]poststarthook/priority-and-fairness-config-producer ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]poststarthook/bootstrap-controller ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]poststarthook/start-kube-aggregator-informers ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]poststarthook/apiservice-status-local-available-controller ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]poststarthook/apiservice-status-remote-available-controller ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]poststarthook/apiservice-registration-controller ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]poststarthook/apiservice-wait-for-first-sync ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]poststarthook/apiservice-discovery-controller ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]poststarthook/kube-apiserver-autoregistration ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]autoregister-completion ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]poststarthook/apiservice-openapi-controller ok Feb 18 14:55:18 crc kubenswrapper[4896]: [+]poststarthook/apiservice-openapiv3-controller ok Feb 18 14:55:18 crc kubenswrapper[4896]: livez check failed Feb 18 14:55:18 crc kubenswrapper[4896]: I0218 14:55:18.420027 4896 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 14:55:18 crc kubenswrapper[4896]: I0218 14:55:18.857572 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 07:04:58.031372394 +0000 UTC Feb 18 14:55:19 crc kubenswrapper[4896]: I0218 14:55:19.858618 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 08:06:07.570526677 +0000 UTC Feb 18 14:55:20 crc kubenswrapper[4896]: I0218 14:55:20.081467 4896 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 18 14:55:20 crc kubenswrapper[4896]: I0218 14:55:20.859695 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 07:12:13.38708111 +0000 UTC Feb 18 14:55:21 crc kubenswrapper[4896]: I0218 14:55:21.860326 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 04:27:48.04400467 +0000 UTC Feb 18 14:55:22 crc kubenswrapper[4896]: E0218 14:55:22.044158 4896 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 18 14:55:22 crc kubenswrapper[4896]: E0218 14:55:22.298715 4896 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.300544 4896 trace.go:236] Trace[1843920511]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (18-Feb-2026 14:55:10.741) (total time: 11558ms): Feb 18 14:55:22 crc kubenswrapper[4896]: Trace[1843920511]: ---"Objects listed" error: 11558ms (14:55:22.300) Feb 18 14:55:22 crc kubenswrapper[4896]: Trace[1843920511]: [11.558536717s] [11.558536717s] END Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.300563 4896 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.302638 4896 trace.go:236] Trace[272655712]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (18-Feb-2026 14:55:10.240) (total time: 12062ms): Feb 18 14:55:22 crc kubenswrapper[4896]: Trace[272655712]: ---"Objects listed" error: 12062ms (14:55:22.302) Feb 18 14:55:22 crc kubenswrapper[4896]: Trace[272655712]: [12.062570025s] [12.062570025s] END Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.302656 4896 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.302978 4896 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.303129 4896 trace.go:236] Trace[1374174940]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (18-Feb-2026 14:55:09.176) (total time: 13126ms): Feb 18 14:55:22 crc kubenswrapper[4896]: Trace[1374174940]: ---"Objects listed" error: 13126ms (14:55:22.303) Feb 18 14:55:22 crc kubenswrapper[4896]: Trace[1374174940]: [13.126056868s] [13.126056868s] END Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.303165 4896 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 18 14:55:22 crc kubenswrapper[4896]: E0218 14:55:22.303369 4896 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.317652 4896 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.341262 4896 csr.go:261] certificate signing request csr-rn9m9 is approved, waiting to be issued Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.353139 4896 csr.go:257] certificate signing request csr-rn9m9 is issued Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.353555 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.368809 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.824953 4896 apiserver.go:52] "Watching apiserver" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.828565 4896 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.828845 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-etcd/etcd-crc"] Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.829143 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.829278 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.829333 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.829634 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.829716 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.829869 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:55:22 crc kubenswrapper[4896]: E0218 14:55:22.829921 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:55:22 crc kubenswrapper[4896]: E0218 14:55:22.830033 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:55:22 crc kubenswrapper[4896]: E0218 14:55:22.830075 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.830388 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.834623 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.834757 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.834773 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.837171 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.838984 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.839169 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.839265 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.839410 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.847983 4896 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.860864 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 09:34:30.593702791 +0000 UTC Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.863018 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.864888 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.870143 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.875748 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.878378 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.888171 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.902972 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.905845 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.905949 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.906023 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.906087 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.906147 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.906223 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.906310 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.906380 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.906447 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.906508 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.906573 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.906637 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.906702 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.906800 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.906874 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.906255 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.906287 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.906426 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.906586 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.906761 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.906917 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.907117 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.907181 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.907257 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.907336 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.907401 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.907463 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.907523 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.907582 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.907653 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.907721 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.907782 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.907845 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.907907 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.907990 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.908062 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.908126 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.908217 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.908317 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.908396 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.908462 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.908522 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.908592 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.908664 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.907141 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.907305 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.907365 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.907371 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.907555 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.907577 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.907624 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.907807 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.907813 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.907863 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.908026 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.908111 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.908145 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.908215 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.908370 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.908387 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.908421 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.908548 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.908589 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.908663 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.909490 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.909484 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.909615 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.910072 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.910085 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.910079 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.910310 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.910354 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.910470 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.910539 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.910599 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.910637 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.910660 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.910874 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.910917 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911001 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911063 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911065 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911092 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911131 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911159 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911178 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911202 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911099 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911481 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911533 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911550 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911567 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911584 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911600 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911618 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911635 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911653 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911647 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911671 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911740 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911764 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911783 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911798 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911813 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911828 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911823 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911849 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911876 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911894 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911912 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911929 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911944 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911975 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.911993 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.912012 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.912027 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.912042 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.912057 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.912072 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.912088 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.912108 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.912124 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.912121 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.912143 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.912331 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.912281 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.912364 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.912402 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.912450 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.912491 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.912508 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.912533 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.912581 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.912624 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.912662 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.912703 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.912713 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.912742 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.912783 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.912820 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.912866 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.912910 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.912947 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.912956 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.912984 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913024 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913028 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913070 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913090 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913109 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913126 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913148 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913165 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913183 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913203 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913221 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913257 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913278 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913297 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913316 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913333 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913357 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913374 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913391 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913408 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913424 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913441 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913459 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913476 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913494 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913509 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913525 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913543 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913559 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913575 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913590 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913605 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913621 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913638 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913653 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913670 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913686 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913701 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913717 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913732 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913750 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913766 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913782 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913853 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913870 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913887 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913903 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913919 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913935 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913949 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913992 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914012 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914029 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914046 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914064 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914082 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914101 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914117 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914135 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914153 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914171 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914187 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914202 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914217 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914248 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914266 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914282 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914298 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914329 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914345 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914363 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914379 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914398 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914415 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914431 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914447 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914465 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914481 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914498 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914513 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914528 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914545 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914561 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914577 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914597 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914616 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914632 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914649 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914664 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914683 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914701 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914718 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915201 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915243 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915266 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915288 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915308 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915355 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915378 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915399 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915420 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915441 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915461 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915489 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915508 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915528 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915547 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915572 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915599 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915637 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915655 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915729 4896 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915740 4896 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915751 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915762 4896 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915773 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915784 4896 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915796 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915806 4896 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915825 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915838 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915849 4896 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915859 4896 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915868 4896 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915877 4896 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915887 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915897 4896 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915907 4896 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915917 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915927 4896 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915938 4896 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915949 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915958 4896 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915967 4896 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915976 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915989 4896 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.916001 4896 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.916014 4896 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.916044 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.916057 4896 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.916066 4896 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.916079 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.916089 4896 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.916098 4896 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.916108 4896 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.916117 4896 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.916127 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.916136 4896 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.916145 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.916155 4896 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.916164 4896 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.916174 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.916183 4896 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.916192 4896 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.916201 4896 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.916210 4896 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.916219 4896 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.916228 4896 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.917010 4896 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.921224 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.921591 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.921695 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.925268 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913036 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913543 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913610 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.913962 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914009 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914267 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914342 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914586 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914695 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.914866 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915019 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915142 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915224 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915764 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.915788 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.916028 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.916472 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.916740 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.916745 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.916912 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.917021 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.917461 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.917763 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.917867 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.918034 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.918049 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.918057 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.917488 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.918330 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.918351 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.918365 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.918544 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.918530 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.918904 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.918969 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.920670 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.920887 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: E0218 14:55:22.921110 4896 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.921376 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.921514 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.921623 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.921818 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.921856 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.921945 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: E0218 14:55:22.921996 4896 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.922132 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.922325 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.922491 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.922584 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.922680 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.923070 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.923140 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.923577 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.923715 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.923789 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.924081 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.924604 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.924896 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.925179 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.927047 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.927082 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.927579 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.927685 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.927924 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.928701 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.928849 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.929303 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.929526 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.929787 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.930153 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.932513 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.932858 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.933352 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.933577 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.933870 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.934169 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.934179 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.935336 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.935338 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.935628 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.935708 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.935701 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.935756 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: E0218 14:55:22.935857 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:55:23.435826543 +0000 UTC m=+22.365160766 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:55:22 crc kubenswrapper[4896]: E0218 14:55:22.945316 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 14:55:23.445278614 +0000 UTC m=+22.374613017 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.945342 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: E0218 14:55:22.945359 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 14:55:23.445346775 +0000 UTC m=+22.374681208 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.936134 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.936247 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.936452 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.936804 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.937324 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.937399 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.938338 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.938451 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.938444 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.938740 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.938841 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.939328 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.939866 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.939964 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.940023 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.940315 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.940549 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.940672 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.941874 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.942116 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.942192 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.942399 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.942437 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.942822 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.943109 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.943148 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.944025 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.944059 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.944134 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.936013 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.948229 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.948640 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.948784 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.948845 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: E0218 14:55:22.959889 4896 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 14:55:22 crc kubenswrapper[4896]: E0218 14:55:22.959941 4896 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 14:55:22 crc kubenswrapper[4896]: E0218 14:55:22.959960 4896 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 14:55:22 crc kubenswrapper[4896]: E0218 14:55:22.960072 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-18 14:55:23.460033124 +0000 UTC m=+22.389367367 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.960516 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.961251 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.961517 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.963602 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.964872 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.965085 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.965754 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.968411 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.969319 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.969962 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.974361 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.974427 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.978615 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.978866 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.985607 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 14:55:22 crc kubenswrapper[4896]: E0218 14:55:22.985873 4896 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 14:55:22 crc kubenswrapper[4896]: E0218 14:55:22.985899 4896 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 14:55:22 crc kubenswrapper[4896]: E0218 14:55:22.985914 4896 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 14:55:22 crc kubenswrapper[4896]: E0218 14:55:22.985977 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-18 14:55:23.48595488 +0000 UTC m=+22.415289093 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.991761 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.992449 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.992957 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.993352 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.994648 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.995384 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.995402 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.995539 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.995601 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.995758 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.997654 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.997896 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:22 crc kubenswrapper[4896]: I0218 14:55:22.998115 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.001018 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.005833 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.006407 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.006555 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.006656 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.007888 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.009614 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.015856 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017185 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017559 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017587 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017635 4896 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017647 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017656 4896 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017665 4896 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017673 4896 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017682 4896 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017691 4896 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017699 4896 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017707 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017716 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017724 4896 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017732 4896 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017741 4896 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017749 4896 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017759 4896 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017767 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017777 4896 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017786 4896 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017796 4896 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017804 4896 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017813 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017821 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017829 4896 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017837 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017844 4896 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017854 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017863 4896 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017873 4896 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017884 4896 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017895 4896 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017907 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017919 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017930 4896 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017940 4896 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017950 4896 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017961 4896 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017972 4896 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.017981 4896 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018034 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018048 4896 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018063 4896 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018074 4896 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018085 4896 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018097 4896 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018107 4896 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018117 4896 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018126 4896 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018135 4896 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018145 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018155 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018165 4896 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018174 4896 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018184 4896 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018194 4896 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018206 4896 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018218 4896 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018217 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018228 4896 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018285 4896 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018299 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018311 4896 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018323 4896 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018335 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018346 4896 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018356 4896 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018366 4896 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018377 4896 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018388 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018401 4896 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018414 4896 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018417 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018426 4896 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018470 4896 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018484 4896 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018510 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018524 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018534 4896 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018542 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018550 4896 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018558 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018567 4896 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018576 4896 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018586 4896 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018594 4896 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018603 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018612 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018621 4896 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018629 4896 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018638 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018646 4896 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018654 4896 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018663 4896 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018670 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018679 4896 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018687 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018694 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018702 4896 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018710 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018719 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018726 4896 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018734 4896 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018742 4896 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018750 4896 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018759 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018768 4896 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018777 4896 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018785 4896 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018809 4896 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018820 4896 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018829 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018838 4896 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018846 4896 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018854 4896 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018867 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018875 4896 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018884 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018892 4896 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018899 4896 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018908 4896 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018916 4896 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018924 4896 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018932 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018941 4896 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018949 4896 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018957 4896 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018965 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018973 4896 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.018981 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.019006 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.019014 4896 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.019022 4896 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.019030 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.019040 4896 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.019047 4896 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.019055 4896 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.019063 4896 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.019072 4896 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.019081 4896 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.019091 4896 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.019105 4896 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.019121 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.019136 4896 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.019146 4896 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.019156 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.019164 4896 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.019173 4896 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.019214 4896 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.019225 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.022642 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.024973 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.047161 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 14:55:23 crc kubenswrapper[4896]: E0218 14:55:23.055159 4896 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 14:55:23 crc kubenswrapper[4896]: E0218 14:55:23.058858 4896 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.060847 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.120362 4896 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.145274 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.159228 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 18 14:55:23 crc kubenswrapper[4896]: W0218 14:55:23.159761 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-a56a22d8f7c3ac541c5f8d1e9cd27153e236fae85768c022dcd59dbc004152bd WatchSource:0}: Error finding container a56a22d8f7c3ac541c5f8d1e9cd27153e236fae85768c022dcd59dbc004152bd: Status 404 returned error can't find the container with id a56a22d8f7c3ac541c5f8d1e9cd27153e236fae85768c022dcd59dbc004152bd Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.165015 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 18 14:55:23 crc kubenswrapper[4896]: W0218 14:55:23.171653 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-abf8e760c6d9e49bd5bc28adfb1746d800b338039a68b51938228038367610ad WatchSource:0}: Error finding container abf8e760c6d9e49bd5bc28adfb1746d800b338039a68b51938228038367610ad: Status 404 returned error can't find the container with id abf8e760c6d9e49bd5bc28adfb1746d800b338039a68b51938228038367610ad Feb 18 14:55:23 crc kubenswrapper[4896]: W0218 14:55:23.179846 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-d76cb58c0bfcc50c66fa36f3d8abb04e21d1ffdacb0a99b74a85be724c426582 WatchSource:0}: Error finding container d76cb58c0bfcc50c66fa36f3d8abb04e21d1ffdacb0a99b74a85be724c426582: Status 404 returned error can't find the container with id d76cb58c0bfcc50c66fa36f3d8abb04e21d1ffdacb0a99b74a85be724c426582 Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.354296 4896 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-18 14:50:22 +0000 UTC, rotation deadline is 2026-12-31 12:22:00.648064924 +0000 UTC Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.354364 4896 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7581h26m37.293704255s for next certificate rotation Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.420296 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.420969 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.424181 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.433265 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.435368 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.445130 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.455049 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.477016 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.509221 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.523937 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.524052 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.524086 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:55:23 crc kubenswrapper[4896]: E0218 14:55:23.524173 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:55:24.524138782 +0000 UTC m=+23.453472995 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:55:23 crc kubenswrapper[4896]: E0218 14:55:23.524177 4896 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.524213 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.524255 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:55:23 crc kubenswrapper[4896]: E0218 14:55:23.524272 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 14:55:24.524263506 +0000 UTC m=+23.453597719 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 14:55:23 crc kubenswrapper[4896]: E0218 14:55:23.524299 4896 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 14:55:23 crc kubenswrapper[4896]: E0218 14:55:23.524330 4896 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 14:55:23 crc kubenswrapper[4896]: E0218 14:55:23.524341 4896 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 14:55:23 crc kubenswrapper[4896]: E0218 14:55:23.524377 4896 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 14:55:23 crc kubenswrapper[4896]: E0218 14:55:23.524385 4896 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 14:55:23 crc kubenswrapper[4896]: E0218 14:55:23.524419 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-18 14:55:24.524399669 +0000 UTC m=+23.453733882 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 14:55:23 crc kubenswrapper[4896]: E0218 14:55:23.524487 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 14:55:24.52442684 +0000 UTC m=+23.453761233 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 14:55:23 crc kubenswrapper[4896]: E0218 14:55:23.524396 4896 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 14:55:23 crc kubenswrapper[4896]: E0218 14:55:23.524513 4896 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 14:55:23 crc kubenswrapper[4896]: E0218 14:55:23.524552 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-18 14:55:24.524543443 +0000 UTC m=+23.453877656 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.527834 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.547882 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.567665 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.580152 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.599708 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.613272 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71647042503d25307e9924a70b6a802a8d0d0ff4a43eff413dbfa517c0f4eb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:16Z\\\",\\\"message\\\":\\\"W0218 14:55:05.294340 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 14:55:05.294825 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771426505 cert, and key in /tmp/serving-cert-175718111/serving-signer.crt, /tmp/serving-cert-175718111/serving-signer.key\\\\nI0218 14:55:05.618615 1 observer_polling.go:159] Starting file observer\\\\nW0218 14:55:05.621773 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 14:55:05.621937 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:05.623210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-175718111/tls.crt::/tmp/serving-cert-175718111/tls.key\\\\\\\"\\\\nF0218 14:55:16.036550 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.629393 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.639726 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.650969 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.705737 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.718967 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.731622 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.861869 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 20:17:13.252387771 +0000 UTC Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.931679 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.932670 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.933714 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.934588 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.935459 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.936188 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.937034 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.937829 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.938694 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.939518 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.940290 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.944026 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.944822 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.945653 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.946908 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.947691 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.949147 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.949681 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.950636 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.952031 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.952699 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.953981 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.954638 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.955584 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.956754 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.957608 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.959055 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.959772 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.961060 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.961916 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.962566 4896 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.963363 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.965703 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.966842 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.967687 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.969730 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.971129 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.971856 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.973324 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.974217 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.974896 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.976321 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.977346 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.977927 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.978741 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.979276 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.980110 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.980858 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.981658 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.982073 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.982593 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.983681 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.984393 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Feb 18 14:55:23 crc kubenswrapper[4896]: I0218 14:55:23.985398 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.019750 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-bxlt7"] Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.020124 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-bxlt7" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.021831 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.022123 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.022292 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.033829 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.045522 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.051761 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.052670 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.053960 4896 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23" exitCode=255 Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.054020 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23"} Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.054069 4896 scope.go:117] "RemoveContainer" containerID="71647042503d25307e9924a70b6a802a8d0d0ff4a43eff413dbfa517c0f4eb97" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.054646 4896 scope.go:117] "RemoveContainer" containerID="fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23" Feb 18 14:55:24 crc kubenswrapper[4896]: E0218 14:55:24.054833 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.057140 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887"} Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.057182 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd"} Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.057197 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"d76cb58c0bfcc50c66fa36f3d8abb04e21d1ffdacb0a99b74a85be724c426582"} Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.058630 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"abf8e760c6d9e49bd5bc28adfb1746d800b338039a68b51938228038367610ad"} Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.060254 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73"} Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.060285 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"a56a22d8f7c3ac541c5f8d1e9cd27153e236fae85768c022dcd59dbc004152bd"} Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.067450 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.106686 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.123899 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.127486 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3c76e871-21e3-4654-9fbb-9770bf54ee0b-hosts-file\") pod \"node-resolver-bxlt7\" (UID: \"3c76e871-21e3-4654-9fbb-9770bf54ee0b\") " pod="openshift-dns/node-resolver-bxlt7" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.127602 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2bpm\" (UniqueName: \"kubernetes.io/projected/3c76e871-21e3-4654-9fbb-9770bf54ee0b-kube-api-access-h2bpm\") pod \"node-resolver-bxlt7\" (UID: \"3c76e871-21e3-4654-9fbb-9770bf54ee0b\") " pod="openshift-dns/node-resolver-bxlt7" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.184451 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.208501 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.228582 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2bpm\" (UniqueName: \"kubernetes.io/projected/3c76e871-21e3-4654-9fbb-9770bf54ee0b-kube-api-access-h2bpm\") pod \"node-resolver-bxlt7\" (UID: \"3c76e871-21e3-4654-9fbb-9770bf54ee0b\") " pod="openshift-dns/node-resolver-bxlt7" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.228637 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3c76e871-21e3-4654-9fbb-9770bf54ee0b-hosts-file\") pod \"node-resolver-bxlt7\" (UID: \"3c76e871-21e3-4654-9fbb-9770bf54ee0b\") " pod="openshift-dns/node-resolver-bxlt7" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.228757 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/3c76e871-21e3-4654-9fbb-9770bf54ee0b-hosts-file\") pod \"node-resolver-bxlt7\" (UID: \"3c76e871-21e3-4654-9fbb-9770bf54ee0b\") " pod="openshift-dns/node-resolver-bxlt7" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.234208 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.244867 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2bpm\" (UniqueName: \"kubernetes.io/projected/3c76e871-21e3-4654-9fbb-9770bf54ee0b-kube-api-access-h2bpm\") pod \"node-resolver-bxlt7\" (UID: \"3c76e871-21e3-4654-9fbb-9770bf54ee0b\") " pod="openshift-dns/node-resolver-bxlt7" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.245321 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71647042503d25307e9924a70b6a802a8d0d0ff4a43eff413dbfa517c0f4eb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:16Z\\\",\\\"message\\\":\\\"W0218 14:55:05.294340 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 14:55:05.294825 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771426505 cert, and key in /tmp/serving-cert-175718111/serving-signer.crt, /tmp/serving-cert-175718111/serving-signer.key\\\\nI0218 14:55:05.618615 1 observer_polling.go:159] Starting file observer\\\\nW0218 14:55:05.621773 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 14:55:05.621937 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:05.623210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-175718111/tls.crt::/tmp/serving-cert-175718111/tls.key\\\\\\\"\\\\nF0218 14:55:16.036550 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.261685 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.273708 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.292908 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.307898 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71647042503d25307e9924a70b6a802a8d0d0ff4a43eff413dbfa517c0f4eb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:16Z\\\",\\\"message\\\":\\\"W0218 14:55:05.294340 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 14:55:05.294825 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771426505 cert, and key in /tmp/serving-cert-175718111/serving-signer.crt, /tmp/serving-cert-175718111/serving-signer.key\\\\nI0218 14:55:05.618615 1 observer_polling.go:159] Starting file observer\\\\nW0218 14:55:05.621773 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 14:55:05.621937 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:05.623210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-175718111/tls.crt::/tmp/serving-cert-175718111/tls.key\\\\\\\"\\\\nF0218 14:55:16.036550 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.323916 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.331622 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-bxlt7" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.338542 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.353529 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:24 crc kubenswrapper[4896]: W0218 14:55:24.353712 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c76e871_21e3_4654_9fbb_9770bf54ee0b.slice/crio-a3fe441314cd0009251e8e687a4d53f66b56e612a8bc21628abf0fa635d25c05 WatchSource:0}: Error finding container a3fe441314cd0009251e8e687a4d53f66b56e612a8bc21628abf0fa635d25c05: Status 404 returned error can't find the container with id a3fe441314cd0009251e8e687a4d53f66b56e612a8bc21628abf0fa635d25c05 Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.367657 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.382843 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.393133 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.405128 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.530785 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.530857 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:55:24 crc kubenswrapper[4896]: E0218 14:55:24.530875 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:55:26.530854623 +0000 UTC m=+25.460188836 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.530903 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.530936 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.530962 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:55:24 crc kubenswrapper[4896]: E0218 14:55:24.530942 4896 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 14:55:24 crc kubenswrapper[4896]: E0218 14:55:24.531014 4896 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 14:55:24 crc kubenswrapper[4896]: E0218 14:55:24.531033 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 14:55:26.531025117 +0000 UTC m=+25.460359330 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 14:55:24 crc kubenswrapper[4896]: E0218 14:55:24.531047 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 14:55:26.531038658 +0000 UTC m=+25.460372871 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 14:55:24 crc kubenswrapper[4896]: E0218 14:55:24.530997 4896 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 14:55:24 crc kubenswrapper[4896]: E0218 14:55:24.531067 4896 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 14:55:24 crc kubenswrapper[4896]: E0218 14:55:24.531079 4896 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 14:55:24 crc kubenswrapper[4896]: E0218 14:55:24.531107 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-18 14:55:26.531097699 +0000 UTC m=+25.460431912 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 14:55:24 crc kubenswrapper[4896]: E0218 14:55:24.531132 4896 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 14:55:24 crc kubenswrapper[4896]: E0218 14:55:24.531169 4896 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 14:55:24 crc kubenswrapper[4896]: E0218 14:55:24.531181 4896 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 14:55:24 crc kubenswrapper[4896]: E0218 14:55:24.531245 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-18 14:55:26.531214782 +0000 UTC m=+25.460548995 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.862885 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 17:48:47.812624019 +0000 UTC Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.910888 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-24rq5"] Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.911211 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.911816 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-fmg2f"] Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.911979 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-fmg2f" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.912757 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.913147 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.913453 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-scgl4"] Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.913504 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.913719 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.914020 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-scgl4" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.914151 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.914316 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.917307 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.917357 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.917451 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.918506 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.920519 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.920690 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.927455 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.927471 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.927455 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:55:24 crc kubenswrapper[4896]: E0218 14:55:24.927569 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:55:24 crc kubenswrapper[4896]: E0218 14:55:24.927676 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:55:24 crc kubenswrapper[4896]: E0218 14:55:24.927749 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.940026 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.953577 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71647042503d25307e9924a70b6a802a8d0d0ff4a43eff413dbfa517c0f4eb97\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:16Z\\\",\\\"message\\\":\\\"W0218 14:55:05.294340 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI0218 14:55:05.294825 1 crypto.go:601] Generating new CA for check-endpoints-signer@1771426505 cert, and key in /tmp/serving-cert-175718111/serving-signer.crt, /tmp/serving-cert-175718111/serving-signer.key\\\\nI0218 14:55:05.618615 1 observer_polling.go:159] Starting file observer\\\\nW0218 14:55:05.621773 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI0218 14:55:05.621937 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:05.623210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-175718111/tls.crt::/tmp/serving-cert-175718111/tls.key\\\\\\\"\\\\nF0218 14:55:16.036550 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.965008 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.974605 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:24 crc kubenswrapper[4896]: I0218 14:55:24.988432 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.005351 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.018889 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.032026 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.034187 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e2efa13d-b712-4f54-967a-d948be1fce0e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-scgl4\" (UID: \"e2efa13d-b712-4f54-967a-d948be1fce0e\") " pod="openshift-multus/multus-additional-cni-plugins-scgl4" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.034255 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e2efa13d-b712-4f54-967a-d948be1fce0e-cni-binary-copy\") pod \"multus-additional-cni-plugins-scgl4\" (UID: \"e2efa13d-b712-4f54-967a-d948be1fce0e\") " pod="openshift-multus/multus-additional-cni-plugins-scgl4" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.034278 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e2efa13d-b712-4f54-967a-d948be1fce0e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-scgl4\" (UID: \"e2efa13d-b712-4f54-967a-d948be1fce0e\") " pod="openshift-multus/multus-additional-cni-plugins-scgl4" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.034316 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-system-cni-dir\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.034337 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-hostroot\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.034358 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-etc-kubernetes\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.034377 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e2efa13d-b712-4f54-967a-d948be1fce0e-cnibin\") pod \"multus-additional-cni-plugins-scgl4\" (UID: \"e2efa13d-b712-4f54-967a-d948be1fce0e\") " pod="openshift-multus/multus-additional-cni-plugins-scgl4" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.034398 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-multus-socket-dir-parent\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.034417 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-multus-daemon-config\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.034435 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-os-release\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.034455 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-cnibin\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.034475 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-multus-conf-dir\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.034494 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rn6c4\" (UniqueName: \"kubernetes.io/projected/e2efa13d-b712-4f54-967a-d948be1fce0e-kube-api-access-rn6c4\") pod \"multus-additional-cni-plugins-scgl4\" (UID: \"e2efa13d-b712-4f54-967a-d948be1fce0e\") " pod="openshift-multus/multus-additional-cni-plugins-scgl4" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.034522 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8503734a-8111-4fab-beda-1f0f3b59615b-rootfs\") pod \"machine-config-daemon-24rq5\" (UID: \"8503734a-8111-4fab-beda-1f0f3b59615b\") " pod="openshift-machine-config-operator/machine-config-daemon-24rq5" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.034542 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-host-run-k8s-cni-cncf-io\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.034563 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-host-run-multus-certs\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.034585 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e2efa13d-b712-4f54-967a-d948be1fce0e-os-release\") pod \"multus-additional-cni-plugins-scgl4\" (UID: \"e2efa13d-b712-4f54-967a-d948be1fce0e\") " pod="openshift-multus/multus-additional-cni-plugins-scgl4" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.034603 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e2efa13d-b712-4f54-967a-d948be1fce0e-system-cni-dir\") pod \"multus-additional-cni-plugins-scgl4\" (UID: \"e2efa13d-b712-4f54-967a-d948be1fce0e\") " pod="openshift-multus/multus-additional-cni-plugins-scgl4" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.034622 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-cni-binary-copy\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.034641 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-host-run-netns\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.034658 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-host-var-lib-kubelet\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.034677 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5rmg\" (UniqueName: \"kubernetes.io/projected/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-kube-api-access-w5rmg\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.034697 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-host-var-lib-cni-bin\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.034718 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5k2b\" (UniqueName: \"kubernetes.io/projected/8503734a-8111-4fab-beda-1f0f3b59615b-kube-api-access-n5k2b\") pod \"machine-config-daemon-24rq5\" (UID: \"8503734a-8111-4fab-beda-1f0f3b59615b\") " pod="openshift-machine-config-operator/machine-config-daemon-24rq5" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.034738 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8503734a-8111-4fab-beda-1f0f3b59615b-mcd-auth-proxy-config\") pod \"machine-config-daemon-24rq5\" (UID: \"8503734a-8111-4fab-beda-1f0f3b59615b\") " pod="openshift-machine-config-operator/machine-config-daemon-24rq5" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.034821 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-multus-cni-dir\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.034857 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-host-var-lib-cni-multus\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.034899 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8503734a-8111-4fab-beda-1f0f3b59615b-proxy-tls\") pod \"machine-config-daemon-24rq5\" (UID: \"8503734a-8111-4fab-beda-1f0f3b59615b\") " pod="openshift-machine-config-operator/machine-config-daemon-24rq5" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.044857 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.058874 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.063032 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.065890 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-bxlt7" event={"ID":"3c76e871-21e3-4654-9fbb-9770bf54ee0b","Type":"ContainerStarted","Data":"5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f"} Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.065939 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-bxlt7" event={"ID":"3c76e871-21e3-4654-9fbb-9770bf54ee0b","Type":"ContainerStarted","Data":"a3fe441314cd0009251e8e687a4d53f66b56e612a8bc21628abf0fa635d25c05"} Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.066151 4896 scope.go:117] "RemoveContainer" containerID="fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23" Feb 18 14:55:25 crc kubenswrapper[4896]: E0218 14:55:25.066313 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.071011 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.083436 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.096039 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.124307 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.135972 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-host-var-lib-cni-multus\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136006 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8503734a-8111-4fab-beda-1f0f3b59615b-proxy-tls\") pod \"machine-config-daemon-24rq5\" (UID: \"8503734a-8111-4fab-beda-1f0f3b59615b\") " pod="openshift-machine-config-operator/machine-config-daemon-24rq5" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136039 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-multus-cni-dir\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136066 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-host-var-lib-cni-multus\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136077 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e2efa13d-b712-4f54-967a-d948be1fce0e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-scgl4\" (UID: \"e2efa13d-b712-4f54-967a-d948be1fce0e\") " pod="openshift-multus/multus-additional-cni-plugins-scgl4" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136141 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e2efa13d-b712-4f54-967a-d948be1fce0e-cni-binary-copy\") pod \"multus-additional-cni-plugins-scgl4\" (UID: \"e2efa13d-b712-4f54-967a-d948be1fce0e\") " pod="openshift-multus/multus-additional-cni-plugins-scgl4" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136167 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e2efa13d-b712-4f54-967a-d948be1fce0e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-scgl4\" (UID: \"e2efa13d-b712-4f54-967a-d948be1fce0e\") " pod="openshift-multus/multus-additional-cni-plugins-scgl4" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136217 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-system-cni-dir\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136259 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-hostroot\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136280 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-etc-kubernetes\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136280 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-multus-cni-dir\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136302 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e2efa13d-b712-4f54-967a-d948be1fce0e-cnibin\") pod \"multus-additional-cni-plugins-scgl4\" (UID: \"e2efa13d-b712-4f54-967a-d948be1fce0e\") " pod="openshift-multus/multus-additional-cni-plugins-scgl4" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136325 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-hostroot\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136333 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-multus-socket-dir-parent\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136362 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-multus-socket-dir-parent\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136367 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-multus-daemon-config\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136385 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e2efa13d-b712-4f54-967a-d948be1fce0e-cnibin\") pod \"multus-additional-cni-plugins-scgl4\" (UID: \"e2efa13d-b712-4f54-967a-d948be1fce0e\") " pod="openshift-multus/multus-additional-cni-plugins-scgl4" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136398 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-etc-kubernetes\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136394 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-os-release\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136476 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-system-cni-dir\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136509 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-multus-conf-dir\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136545 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rn6c4\" (UniqueName: \"kubernetes.io/projected/e2efa13d-b712-4f54-967a-d948be1fce0e-kube-api-access-rn6c4\") pod \"multus-additional-cni-plugins-scgl4\" (UID: \"e2efa13d-b712-4f54-967a-d948be1fce0e\") " pod="openshift-multus/multus-additional-cni-plugins-scgl4" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136594 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-cnibin\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136624 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8503734a-8111-4fab-beda-1f0f3b59615b-rootfs\") pod \"machine-config-daemon-24rq5\" (UID: \"8503734a-8111-4fab-beda-1f0f3b59615b\") " pod="openshift-machine-config-operator/machine-config-daemon-24rq5" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136641 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-multus-conf-dir\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136651 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-host-run-k8s-cni-cncf-io\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136691 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-cnibin\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136684 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-host-run-k8s-cni-cncf-io\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136721 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-host-run-multus-certs\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136747 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-os-release\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136760 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e2efa13d-b712-4f54-967a-d948be1fce0e-os-release\") pod \"multus-additional-cni-plugins-scgl4\" (UID: \"e2efa13d-b712-4f54-967a-d948be1fce0e\") " pod="openshift-multus/multus-additional-cni-plugins-scgl4" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136777 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/8503734a-8111-4fab-beda-1f0f3b59615b-rootfs\") pod \"machine-config-daemon-24rq5\" (UID: \"8503734a-8111-4fab-beda-1f0f3b59615b\") " pod="openshift-machine-config-operator/machine-config-daemon-24rq5" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136804 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e2efa13d-b712-4f54-967a-d948be1fce0e-os-release\") pod \"multus-additional-cni-plugins-scgl4\" (UID: \"e2efa13d-b712-4f54-967a-d948be1fce0e\") " pod="openshift-multus/multus-additional-cni-plugins-scgl4" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136795 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e2efa13d-b712-4f54-967a-d948be1fce0e-system-cni-dir\") pod \"multus-additional-cni-plugins-scgl4\" (UID: \"e2efa13d-b712-4f54-967a-d948be1fce0e\") " pod="openshift-multus/multus-additional-cni-plugins-scgl4" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136828 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e2efa13d-b712-4f54-967a-d948be1fce0e-system-cni-dir\") pod \"multus-additional-cni-plugins-scgl4\" (UID: \"e2efa13d-b712-4f54-967a-d948be1fce0e\") " pod="openshift-multus/multus-additional-cni-plugins-scgl4" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136823 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-host-run-multus-certs\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136864 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-cni-binary-copy\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136884 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-host-run-netns\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136902 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-host-var-lib-kubelet\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136919 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5rmg\" (UniqueName: \"kubernetes.io/projected/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-kube-api-access-w5rmg\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136938 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-host-var-lib-cni-bin\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136948 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-host-run-netns\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136957 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5k2b\" (UniqueName: \"kubernetes.io/projected/8503734a-8111-4fab-beda-1f0f3b59615b-kube-api-access-n5k2b\") pod \"machine-config-daemon-24rq5\" (UID: \"8503734a-8111-4fab-beda-1f0f3b59615b\") " pod="openshift-machine-config-operator/machine-config-daemon-24rq5" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136980 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-host-var-lib-kubelet\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.137022 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-host-var-lib-cni-bin\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.136997 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8503734a-8111-4fab-beda-1f0f3b59615b-mcd-auth-proxy-config\") pod \"machine-config-daemon-24rq5\" (UID: \"8503734a-8111-4fab-beda-1f0f3b59615b\") " pod="openshift-machine-config-operator/machine-config-daemon-24rq5" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.137084 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e2efa13d-b712-4f54-967a-d948be1fce0e-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-scgl4\" (UID: \"e2efa13d-b712-4f54-967a-d948be1fce0e\") " pod="openshift-multus/multus-additional-cni-plugins-scgl4" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.137366 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-multus-daemon-config\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.137456 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-cni-binary-copy\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.137616 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e2efa13d-b712-4f54-967a-d948be1fce0e-cni-binary-copy\") pod \"multus-additional-cni-plugins-scgl4\" (UID: \"e2efa13d-b712-4f54-967a-d948be1fce0e\") " pod="openshift-multus/multus-additional-cni-plugins-scgl4" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.137973 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8503734a-8111-4fab-beda-1f0f3b59615b-mcd-auth-proxy-config\") pod \"machine-config-daemon-24rq5\" (UID: \"8503734a-8111-4fab-beda-1f0f3b59615b\") " pod="openshift-machine-config-operator/machine-config-daemon-24rq5" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.138098 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e2efa13d-b712-4f54-967a-d948be1fce0e-tuning-conf-dir\") pod \"multus-additional-cni-plugins-scgl4\" (UID: \"e2efa13d-b712-4f54-967a-d948be1fce0e\") " pod="openshift-multus/multus-additional-cni-plugins-scgl4" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.139041 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.141798 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8503734a-8111-4fab-beda-1f0f3b59615b-proxy-tls\") pod \"machine-config-daemon-24rq5\" (UID: \"8503734a-8111-4fab-beda-1f0f3b59615b\") " pod="openshift-machine-config-operator/machine-config-daemon-24rq5" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.153987 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5k2b\" (UniqueName: \"kubernetes.io/projected/8503734a-8111-4fab-beda-1f0f3b59615b-kube-api-access-n5k2b\") pod \"machine-config-daemon-24rq5\" (UID: \"8503734a-8111-4fab-beda-1f0f3b59615b\") " pod="openshift-machine-config-operator/machine-config-daemon-24rq5" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.156594 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5rmg\" (UniqueName: \"kubernetes.io/projected/c5ba4981-ee97-4d71-b48f-b5d12bd5911c-kube-api-access-w5rmg\") pod \"multus-fmg2f\" (UID: \"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\") " pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.158804 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.165846 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rn6c4\" (UniqueName: \"kubernetes.io/projected/e2efa13d-b712-4f54-967a-d948be1fce0e-kube-api-access-rn6c4\") pod \"multus-additional-cni-plugins-scgl4\" (UID: \"e2efa13d-b712-4f54-967a-d948be1fce0e\") " pod="openshift-multus/multus-additional-cni-plugins-scgl4" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.176164 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.192497 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.206316 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.223618 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.225502 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: W0218 14:55:25.234647 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8503734a_8111_4fab_beda_1f0f3b59615b.slice/crio-3a3d0fd3116a7092104891a039cd753cf9fb53800bc11e8d6cd0ef220522d021 WatchSource:0}: Error finding container 3a3d0fd3116a7092104891a039cd753cf9fb53800bc11e8d6cd0ef220522d021: Status 404 returned error can't find the container with id 3a3d0fd3116a7092104891a039cd753cf9fb53800bc11e8d6cd0ef220522d021 Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.236165 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-fmg2f" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.240163 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.245304 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-scgl4" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.256652 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: W0218 14:55:25.267764 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode2efa13d_b712_4f54_967a_d948be1fce0e.slice/crio-2069851193e19fa3fa04be647eb24cbbeb897594e90354e520c61a5db3ada1ce WatchSource:0}: Error finding container 2069851193e19fa3fa04be647eb24cbbeb897594e90354e520c61a5db3ada1ce: Status 404 returned error can't find the container with id 2069851193e19fa3fa04be647eb24cbbeb897594e90354e520c61a5db3ada1ce Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.271217 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.289667 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.300751 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-w862c"] Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.301772 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.308717 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.309299 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.309534 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.309890 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.309931 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.310279 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.311524 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.324179 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.344612 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.359008 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.373740 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.387872 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.405819 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.419248 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.432388 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.439212 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-slash\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.439306 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-var-lib-openvswitch\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.439333 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-run-ovn\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.439352 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-run-netns\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.439381 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-run-systemd\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.439405 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-cni-bin\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.439426 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/449d40af-fe23-4350-b66f-de4ce6614177-ovnkube-config\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.439446 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/449d40af-fe23-4350-b66f-de4ce6614177-ovnkube-script-lib\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.439471 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-systemd-units\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.439490 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/449d40af-fe23-4350-b66f-de4ce6614177-ovn-node-metrics-cert\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.439508 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zjjt\" (UniqueName: \"kubernetes.io/projected/449d40af-fe23-4350-b66f-de4ce6614177-kube-api-access-9zjjt\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.439528 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-kubelet\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.439556 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/449d40af-fe23-4350-b66f-de4ce6614177-env-overrides\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.439575 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-node-log\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.439594 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-run-ovn-kubernetes\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.439651 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.439740 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-run-openvswitch\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.439788 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-cni-netd\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.439845 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-etc-openvswitch\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.439886 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-log-socket\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.454453 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.466280 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.480328 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.490347 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.510250 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.522312 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541147 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/449d40af-fe23-4350-b66f-de4ce6614177-ovn-node-metrics-cert\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541197 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zjjt\" (UniqueName: \"kubernetes.io/projected/449d40af-fe23-4350-b66f-de4ce6614177-kube-api-access-9zjjt\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541212 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-kubelet\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541250 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/449d40af-fe23-4350-b66f-de4ce6614177-env-overrides\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541266 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-node-log\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541291 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-run-ovn-kubernetes\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541309 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541326 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-kubelet\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541358 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-run-openvswitch\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541334 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-run-openvswitch\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541388 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-cni-netd\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541411 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-etc-openvswitch\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541432 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-log-socket\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541458 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-run-ovn-kubernetes\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541466 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-slash\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541493 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-slash\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541534 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-var-lib-openvswitch\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541543 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-node-log\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541511 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-var-lib-openvswitch\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541572 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-etc-openvswitch\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541582 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-log-socket\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541595 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541617 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-cni-netd\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541664 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-run-ovn\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541688 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-run-netns\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541726 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-run-systemd\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541747 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-cni-bin\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541750 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-run-netns\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541745 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-run-ovn\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541762 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/449d40af-fe23-4350-b66f-de4ce6614177-ovnkube-config\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541801 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-cni-bin\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541780 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-run-systemd\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541820 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/449d40af-fe23-4350-b66f-de4ce6614177-ovnkube-script-lib\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541856 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-systemd-units\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541915 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-systemd-units\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.541967 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/449d40af-fe23-4350-b66f-de4ce6614177-env-overrides\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.542367 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/449d40af-fe23-4350-b66f-de4ce6614177-ovnkube-config\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.542454 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/449d40af-fe23-4350-b66f-de4ce6614177-ovnkube-script-lib\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.545750 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/449d40af-fe23-4350-b66f-de4ce6614177-ovn-node-metrics-cert\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.558226 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zjjt\" (UniqueName: \"kubernetes.io/projected/449d40af-fe23-4350-b66f-de4ce6614177-kube-api-access-9zjjt\") pod \"ovnkube-node-w862c\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.619229 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:25 crc kubenswrapper[4896]: I0218 14:55:25.863611 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 06:25:43.682351773 +0000 UTC Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.070205 4896 generic.go:334] "Generic (PLEG): container finished" podID="449d40af-fe23-4350-b66f-de4ce6614177" containerID="aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1" exitCode=0 Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.070289 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" event={"ID":"449d40af-fe23-4350-b66f-de4ce6614177","Type":"ContainerDied","Data":"aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1"} Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.070315 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" event={"ID":"449d40af-fe23-4350-b66f-de4ce6614177","Type":"ContainerStarted","Data":"115025435df819683983d7d2d9572a0175d2f56d4641320f8be65a2840ee3941"} Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.071806 4896 generic.go:334] "Generic (PLEG): container finished" podID="e2efa13d-b712-4f54-967a-d948be1fce0e" containerID="68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13" exitCode=0 Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.071876 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" event={"ID":"e2efa13d-b712-4f54-967a-d948be1fce0e","Type":"ContainerDied","Data":"68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13"} Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.071909 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" event={"ID":"e2efa13d-b712-4f54-967a-d948be1fce0e","Type":"ContainerStarted","Data":"2069851193e19fa3fa04be647eb24cbbeb897594e90354e520c61a5db3ada1ce"} Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.073923 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08"} Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.075447 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" event={"ID":"8503734a-8111-4fab-beda-1f0f3b59615b","Type":"ContainerStarted","Data":"fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b"} Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.075513 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" event={"ID":"8503734a-8111-4fab-beda-1f0f3b59615b","Type":"ContainerStarted","Data":"5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc"} Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.075526 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" event={"ID":"8503734a-8111-4fab-beda-1f0f3b59615b","Type":"ContainerStarted","Data":"3a3d0fd3116a7092104891a039cd753cf9fb53800bc11e8d6cd0ef220522d021"} Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.076810 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fmg2f" event={"ID":"c5ba4981-ee97-4d71-b48f-b5d12bd5911c","Type":"ContainerStarted","Data":"434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39"} Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.076847 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fmg2f" event={"ID":"c5ba4981-ee97-4d71-b48f-b5d12bd5911c","Type":"ContainerStarted","Data":"b917ecbe32fa42cf8d8375ec98f5e8a1bc71366b9616130be49de8f80989ee63"} Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.092816 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:26Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.108966 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:26Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.121731 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:26Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.147014 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:26Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.165181 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:26Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.179997 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:26Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.192549 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:26Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.205996 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:26Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.221064 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:26Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.233649 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:26Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.250911 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:26Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.262778 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:26Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.274765 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:26Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.294430 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:26Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.307048 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:26Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.323099 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:26Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.334597 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:26Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.345579 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:26Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.360211 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:26Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.374636 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:26Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.398568 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:26Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.410466 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:26Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.422133 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:26Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.433345 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:26Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.447657 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:26Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.461442 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:26Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.477167 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:26Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.490505 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:26Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.552178 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.552360 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.552423 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.552475 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.552515 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:55:26 crc kubenswrapper[4896]: E0218 14:55:26.552705 4896 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 14:55:26 crc kubenswrapper[4896]: E0218 14:55:26.552733 4896 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 14:55:26 crc kubenswrapper[4896]: E0218 14:55:26.552749 4896 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 14:55:26 crc kubenswrapper[4896]: E0218 14:55:26.552813 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-18 14:55:30.55279055 +0000 UTC m=+29.482124783 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 14:55:26 crc kubenswrapper[4896]: E0218 14:55:26.552919 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:55:30.552905493 +0000 UTC m=+29.482239726 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:55:26 crc kubenswrapper[4896]: E0218 14:55:26.552998 4896 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 14:55:26 crc kubenswrapper[4896]: E0218 14:55:26.553040 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 14:55:30.553027097 +0000 UTC m=+29.482361330 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 14:55:26 crc kubenswrapper[4896]: E0218 14:55:26.553123 4896 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 14:55:26 crc kubenswrapper[4896]: E0218 14:55:26.553143 4896 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 14:55:26 crc kubenswrapper[4896]: E0218 14:55:26.553157 4896 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 14:55:26 crc kubenswrapper[4896]: E0218 14:55:26.553198 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-18 14:55:30.553185441 +0000 UTC m=+29.482519664 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 14:55:26 crc kubenswrapper[4896]: E0218 14:55:26.553281 4896 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 14:55:26 crc kubenswrapper[4896]: E0218 14:55:26.553324 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 14:55:30.553312024 +0000 UTC m=+29.482646247 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.867280 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 20:37:14.416971016 +0000 UTC Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.927426 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.927459 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:55:26 crc kubenswrapper[4896]: I0218 14:55:26.927427 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:55:26 crc kubenswrapper[4896]: E0218 14:55:26.927576 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:55:26 crc kubenswrapper[4896]: E0218 14:55:26.927675 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:55:26 crc kubenswrapper[4896]: E0218 14:55:26.927728 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:55:27 crc kubenswrapper[4896]: I0218 14:55:27.107888 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" event={"ID":"449d40af-fe23-4350-b66f-de4ce6614177","Type":"ContainerStarted","Data":"5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2"} Feb 18 14:55:27 crc kubenswrapper[4896]: I0218 14:55:27.107955 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" event={"ID":"449d40af-fe23-4350-b66f-de4ce6614177","Type":"ContainerStarted","Data":"3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521"} Feb 18 14:55:27 crc kubenswrapper[4896]: I0218 14:55:27.107967 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" event={"ID":"449d40af-fe23-4350-b66f-de4ce6614177","Type":"ContainerStarted","Data":"185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a"} Feb 18 14:55:27 crc kubenswrapper[4896]: I0218 14:55:27.107981 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" event={"ID":"449d40af-fe23-4350-b66f-de4ce6614177","Type":"ContainerStarted","Data":"22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf"} Feb 18 14:55:27 crc kubenswrapper[4896]: I0218 14:55:27.113880 4896 generic.go:334] "Generic (PLEG): container finished" podID="e2efa13d-b712-4f54-967a-d948be1fce0e" containerID="c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e" exitCode=0 Feb 18 14:55:27 crc kubenswrapper[4896]: I0218 14:55:27.113972 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" event={"ID":"e2efa13d-b712-4f54-967a-d948be1fce0e","Type":"ContainerDied","Data":"c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e"} Feb 18 14:55:27 crc kubenswrapper[4896]: I0218 14:55:27.138397 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:27Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:27 crc kubenswrapper[4896]: I0218 14:55:27.152681 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:27Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:27 crc kubenswrapper[4896]: I0218 14:55:27.165892 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:27Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:27 crc kubenswrapper[4896]: I0218 14:55:27.185330 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:27Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:27 crc kubenswrapper[4896]: I0218 14:55:27.199302 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:27Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:27 crc kubenswrapper[4896]: I0218 14:55:27.210590 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:27Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:27 crc kubenswrapper[4896]: I0218 14:55:27.222916 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:27Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:27 crc kubenswrapper[4896]: I0218 14:55:27.233846 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:27Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:27 crc kubenswrapper[4896]: I0218 14:55:27.251836 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:27Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:27 crc kubenswrapper[4896]: I0218 14:55:27.269795 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:27Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:27 crc kubenswrapper[4896]: I0218 14:55:27.282422 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:27Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:27 crc kubenswrapper[4896]: I0218 14:55:27.300529 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:27Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:27 crc kubenswrapper[4896]: I0218 14:55:27.321505 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:27Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:27 crc kubenswrapper[4896]: I0218 14:55:27.337934 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:27Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:27 crc kubenswrapper[4896]: I0218 14:55:27.868258 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 20:52:00.083382877 +0000 UTC Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.120924 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" event={"ID":"449d40af-fe23-4350-b66f-de4ce6614177","Type":"ContainerStarted","Data":"52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b"} Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.120970 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" event={"ID":"449d40af-fe23-4350-b66f-de4ce6614177","Type":"ContainerStarted","Data":"1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc"} Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.122691 4896 generic.go:334] "Generic (PLEG): container finished" podID="e2efa13d-b712-4f54-967a-d948be1fce0e" containerID="ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581" exitCode=0 Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.122722 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" event={"ID":"e2efa13d-b712-4f54-967a-d948be1fce0e","Type":"ContainerDied","Data":"ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581"} Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.137740 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.148674 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.168341 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.194042 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.206366 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.216985 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.229441 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.247297 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.263257 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.278031 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.290268 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.302000 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.313399 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.314635 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-6wprd"] Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.315121 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-6wprd" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.317405 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.318628 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.320007 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.320167 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.336437 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.356476 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.372461 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.386146 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.399189 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.412483 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.424107 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.444937 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.460813 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.474174 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e8896bb9-c6fe-4111-a49a-9dadf23e6855-host\") pod \"node-ca-6wprd\" (UID: \"e8896bb9-c6fe-4111-a49a-9dadf23e6855\") " pod="openshift-image-registry/node-ca-6wprd" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.474297 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5jc4\" (UniqueName: \"kubernetes.io/projected/e8896bb9-c6fe-4111-a49a-9dadf23e6855-kube-api-access-j5jc4\") pod \"node-ca-6wprd\" (UID: \"e8896bb9-c6fe-4111-a49a-9dadf23e6855\") " pod="openshift-image-registry/node-ca-6wprd" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.474344 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e8896bb9-c6fe-4111-a49a-9dadf23e6855-serviceca\") pod \"node-ca-6wprd\" (UID: \"e8896bb9-c6fe-4111-a49a-9dadf23e6855\") " pod="openshift-image-registry/node-ca-6wprd" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.474626 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.488866 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.507580 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.520965 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.532878 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.545159 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.553163 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6wprd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8896bb9-c6fe-4111-a49a-9dadf23e6855\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5jc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6wprd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.576117 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e8896bb9-c6fe-4111-a49a-9dadf23e6855-serviceca\") pod \"node-ca-6wprd\" (UID: \"e8896bb9-c6fe-4111-a49a-9dadf23e6855\") " pod="openshift-image-registry/node-ca-6wprd" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.576276 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e8896bb9-c6fe-4111-a49a-9dadf23e6855-host\") pod \"node-ca-6wprd\" (UID: \"e8896bb9-c6fe-4111-a49a-9dadf23e6855\") " pod="openshift-image-registry/node-ca-6wprd" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.576338 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5jc4\" (UniqueName: \"kubernetes.io/projected/e8896bb9-c6fe-4111-a49a-9dadf23e6855-kube-api-access-j5jc4\") pod \"node-ca-6wprd\" (UID: \"e8896bb9-c6fe-4111-a49a-9dadf23e6855\") " pod="openshift-image-registry/node-ca-6wprd" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.576364 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e8896bb9-c6fe-4111-a49a-9dadf23e6855-host\") pod \"node-ca-6wprd\" (UID: \"e8896bb9-c6fe-4111-a49a-9dadf23e6855\") " pod="openshift-image-registry/node-ca-6wprd" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.577128 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/e8896bb9-c6fe-4111-a49a-9dadf23e6855-serviceca\") pod \"node-ca-6wprd\" (UID: \"e8896bb9-c6fe-4111-a49a-9dadf23e6855\") " pod="openshift-image-registry/node-ca-6wprd" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.593550 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5jc4\" (UniqueName: \"kubernetes.io/projected/e8896bb9-c6fe-4111-a49a-9dadf23e6855-kube-api-access-j5jc4\") pod \"node-ca-6wprd\" (UID: \"e8896bb9-c6fe-4111-a49a-9dadf23e6855\") " pod="openshift-image-registry/node-ca-6wprd" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.628640 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-6wprd" Feb 18 14:55:28 crc kubenswrapper[4896]: W0218 14:55:28.647382 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode8896bb9_c6fe_4111_a49a_9dadf23e6855.slice/crio-89fb0f4caf5c933d519a30897ac61aa70b164524e374dbe0530589586afba627 WatchSource:0}: Error finding container 89fb0f4caf5c933d519a30897ac61aa70b164524e374dbe0530589586afba627: Status 404 returned error can't find the container with id 89fb0f4caf5c933d519a30897ac61aa70b164524e374dbe0530589586afba627 Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.703620 4896 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.705559 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.705594 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.705616 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.705732 4896 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.712190 4896 kubelet_node_status.go:115] "Node was previously registered" node="crc" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.712578 4896 kubelet_node_status.go:79] "Successfully registered node" node="crc" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.713726 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.713765 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.713777 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.713794 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.713810 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:28Z","lastTransitionTime":"2026-02-18T14:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:28 crc kubenswrapper[4896]: E0218 14:55:28.725560 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3d8f65ef-15b5-4cc3-be66-c8dd6cb956e9\\\",\\\"systemUUID\\\":\\\"d255e81f-4350-4697-bbdf-3f95dec57eda\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.729064 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.729101 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.729145 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.729164 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.729175 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:28Z","lastTransitionTime":"2026-02-18T14:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:28 crc kubenswrapper[4896]: E0218 14:55:28.739180 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3d8f65ef-15b5-4cc3-be66-c8dd6cb956e9\\\",\\\"systemUUID\\\":\\\"d255e81f-4350-4697-bbdf-3f95dec57eda\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.742121 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.742184 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.742197 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.742253 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.742267 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:28Z","lastTransitionTime":"2026-02-18T14:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:28 crc kubenswrapper[4896]: E0218 14:55:28.752372 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3d8f65ef-15b5-4cc3-be66-c8dd6cb956e9\\\",\\\"systemUUID\\\":\\\"d255e81f-4350-4697-bbdf-3f95dec57eda\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.755880 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.755903 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.755911 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.755924 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.755951 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:28Z","lastTransitionTime":"2026-02-18T14:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:28 crc kubenswrapper[4896]: E0218 14:55:28.767949 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3d8f65ef-15b5-4cc3-be66-c8dd6cb956e9\\\",\\\"systemUUID\\\":\\\"d255e81f-4350-4697-bbdf-3f95dec57eda\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.771093 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.771122 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.771130 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.771190 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.771200 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:28Z","lastTransitionTime":"2026-02-18T14:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:28 crc kubenswrapper[4896]: E0218 14:55:28.783276 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3d8f65ef-15b5-4cc3-be66-c8dd6cb956e9\\\",\\\"systemUUID\\\":\\\"d255e81f-4350-4697-bbdf-3f95dec57eda\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:28Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:28 crc kubenswrapper[4896]: E0218 14:55:28.783386 4896 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.784581 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.784609 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.784618 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.784631 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.784640 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:28Z","lastTransitionTime":"2026-02-18T14:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.869452 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 19:17:56.88254776 +0000 UTC Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.886580 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.886608 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.886618 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.886635 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.886645 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:28Z","lastTransitionTime":"2026-02-18T14:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.927710 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.927747 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:55:28 crc kubenswrapper[4896]: E0218 14:55:28.927844 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:55:28 crc kubenswrapper[4896]: E0218 14:55:28.927970 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.928050 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:55:28 crc kubenswrapper[4896]: E0218 14:55:28.928121 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.991695 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.991734 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.991745 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.991762 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:28 crc kubenswrapper[4896]: I0218 14:55:28.991771 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:28Z","lastTransitionTime":"2026-02-18T14:55:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.093881 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.093915 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.093923 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.093936 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.093944 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:29Z","lastTransitionTime":"2026-02-18T14:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.128022 4896 generic.go:334] "Generic (PLEG): container finished" podID="e2efa13d-b712-4f54-967a-d948be1fce0e" containerID="6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5" exitCode=0 Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.128144 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" event={"ID":"e2efa13d-b712-4f54-967a-d948be1fce0e","Type":"ContainerDied","Data":"6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5"} Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.129942 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-6wprd" event={"ID":"e8896bb9-c6fe-4111-a49a-9dadf23e6855","Type":"ContainerStarted","Data":"80d2101411bb762554b416ff237ce9ce2455e43c3c5379118322264ffd9dcd94"} Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.130207 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-6wprd" event={"ID":"e8896bb9-c6fe-4111-a49a-9dadf23e6855","Type":"ContainerStarted","Data":"89fb0f4caf5c933d519a30897ac61aa70b164524e374dbe0530589586afba627"} Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.148221 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:29Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.159774 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:29Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.175355 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:29Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.190770 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:29Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.196302 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.196363 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.196379 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.196404 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.196423 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:29Z","lastTransitionTime":"2026-02-18T14:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.204470 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:29Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.224997 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:29Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.248123 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:29Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.264079 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:29Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.281101 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:29Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.299412 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.299453 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.299464 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.299484 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.299497 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:29Z","lastTransitionTime":"2026-02-18T14:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.306312 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:29Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.317324 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6wprd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8896bb9-c6fe-4111-a49a-9dadf23e6855\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5jc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6wprd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:29Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.379952 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:29Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.396224 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:29Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.401747 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.401932 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.402009 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.402069 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.402135 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:29Z","lastTransitionTime":"2026-02-18T14:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.410590 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:29Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.424162 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:29Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.437833 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:29Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.450891 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:29Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.471625 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:29Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.484124 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:29Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.500938 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:29Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.504457 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.504500 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.504512 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.504533 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.504544 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:29Z","lastTransitionTime":"2026-02-18T14:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.513478 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:29Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.523115 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:29Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.537126 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:29Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.548903 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:29Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.562044 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:29Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.578438 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:29Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.589543 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:29Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.607092 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.607130 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.607139 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.607076 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:29Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.607153 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.607254 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:29Z","lastTransitionTime":"2026-02-18T14:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.617323 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6wprd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8896bb9-c6fe-4111-a49a-9dadf23e6855\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80d2101411bb762554b416ff237ce9ce2455e43c3c5379118322264ffd9dcd94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5jc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6wprd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:29Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.628855 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:29Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.710534 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.710573 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.710584 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.710598 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.710608 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:29Z","lastTransitionTime":"2026-02-18T14:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.812748 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.812784 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.812793 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.812805 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.812814 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:29Z","lastTransitionTime":"2026-02-18T14:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.870474 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 15:11:34.10102269 +0000 UTC Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.904032 4896 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.904739 4896 scope.go:117] "RemoveContainer" containerID="fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23" Feb 18 14:55:29 crc kubenswrapper[4896]: E0218 14:55:29.904873 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.915018 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.915055 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.915064 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.915131 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:29 crc kubenswrapper[4896]: I0218 14:55:29.915143 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:29Z","lastTransitionTime":"2026-02-18T14:55:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.017731 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.017842 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.017921 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.017940 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.017951 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:30Z","lastTransitionTime":"2026-02-18T14:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.120550 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.120627 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.120652 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.120684 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.120709 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:30Z","lastTransitionTime":"2026-02-18T14:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.136647 4896 generic.go:334] "Generic (PLEG): container finished" podID="e2efa13d-b712-4f54-967a-d948be1fce0e" containerID="13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e" exitCode=0 Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.136721 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" event={"ID":"e2efa13d-b712-4f54-967a-d948be1fce0e","Type":"ContainerDied","Data":"13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e"} Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.143895 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" event={"ID":"449d40af-fe23-4350-b66f-de4ce6614177","Type":"ContainerStarted","Data":"d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad"} Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.159053 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:30Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.179561 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:30Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.192137 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:30Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.208590 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:30Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.221687 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:30Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.223223 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.223283 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.223298 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.223319 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.223342 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:30Z","lastTransitionTime":"2026-02-18T14:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.235295 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:30Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.260562 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:30Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.272637 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:30Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.285957 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:30Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.295013 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6wprd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8896bb9-c6fe-4111-a49a-9dadf23e6855\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80d2101411bb762554b416ff237ce9ce2455e43c3c5379118322264ffd9dcd94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5jc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6wprd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:30Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.305154 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:30Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.316665 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:30Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.326124 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.326191 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.326205 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.326222 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.326254 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:30Z","lastTransitionTime":"2026-02-18T14:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.326294 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:30Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.343529 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:30Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.356778 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:30Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.429003 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.429038 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.429047 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.429062 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.429073 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:30Z","lastTransitionTime":"2026-02-18T14:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.538690 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.538727 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.538735 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.538749 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.538758 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:30Z","lastTransitionTime":"2026-02-18T14:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.594969 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.595077 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:55:30 crc kubenswrapper[4896]: E0218 14:55:30.595118 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:55:38.595084923 +0000 UTC m=+37.524419136 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.595158 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:55:30 crc kubenswrapper[4896]: E0218 14:55:30.595187 4896 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.595196 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.595216 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:55:30 crc kubenswrapper[4896]: E0218 14:55:30.595202 4896 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 14:55:30 crc kubenswrapper[4896]: E0218 14:55:30.595278 4896 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 14:55:30 crc kubenswrapper[4896]: E0218 14:55:30.595320 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-18 14:55:38.595310999 +0000 UTC m=+37.524645262 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 14:55:30 crc kubenswrapper[4896]: E0218 14:55:30.595251 4896 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 14:55:30 crc kubenswrapper[4896]: E0218 14:55:30.595350 4896 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 14:55:30 crc kubenswrapper[4896]: E0218 14:55:30.595380 4896 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 14:55:30 crc kubenswrapper[4896]: E0218 14:55:30.595381 4896 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 14:55:30 crc kubenswrapper[4896]: E0218 14:55:30.595391 4896 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 14:55:30 crc kubenswrapper[4896]: E0218 14:55:30.595358 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 14:55:38.59534906 +0000 UTC m=+37.524683273 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 14:55:30 crc kubenswrapper[4896]: E0218 14:55:30.595455 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 14:55:38.595440032 +0000 UTC m=+37.524774245 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 14:55:30 crc kubenswrapper[4896]: E0218 14:55:30.595466 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-18 14:55:38.595460653 +0000 UTC m=+37.524794866 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.641098 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.641126 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.641133 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.641146 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.641155 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:30Z","lastTransitionTime":"2026-02-18T14:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.743288 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.743320 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.743328 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.743340 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.743350 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:30Z","lastTransitionTime":"2026-02-18T14:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.847951 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.848026 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.848045 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.848071 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.848086 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:30Z","lastTransitionTime":"2026-02-18T14:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.871490 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 02:40:04.31945811 +0000 UTC Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.927452 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.927521 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.927565 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:55:30 crc kubenswrapper[4896]: E0218 14:55:30.927684 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:55:30 crc kubenswrapper[4896]: E0218 14:55:30.927826 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:55:30 crc kubenswrapper[4896]: E0218 14:55:30.927942 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.951324 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.951375 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.951388 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.951407 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:30 crc kubenswrapper[4896]: I0218 14:55:30.951422 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:30Z","lastTransitionTime":"2026-02-18T14:55:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.053725 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.053762 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.053775 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.053791 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.053803 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:31Z","lastTransitionTime":"2026-02-18T14:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.150270 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" event={"ID":"e2efa13d-b712-4f54-967a-d948be1fce0e","Type":"ContainerDied","Data":"bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a"} Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.150270 4896 generic.go:334] "Generic (PLEG): container finished" podID="e2efa13d-b712-4f54-967a-d948be1fce0e" containerID="bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a" exitCode=0 Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.154968 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.155007 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.155019 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.155037 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.155052 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:31Z","lastTransitionTime":"2026-02-18T14:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.164248 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:31Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.178824 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:31Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.195517 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:31Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.206997 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6wprd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8896bb9-c6fe-4111-a49a-9dadf23e6855\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80d2101411bb762554b416ff237ce9ce2455e43c3c5379118322264ffd9dcd94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5jc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6wprd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:31Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.231853 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:31Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.244312 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:31Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.257428 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:31Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.258740 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.258782 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.258791 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.258835 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.258848 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:31Z","lastTransitionTime":"2026-02-18T14:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.267873 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:31Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.282301 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:31Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.294545 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:31Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.308383 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:31Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.319504 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:31Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.328269 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:31Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.338657 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:31Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.354657 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:31Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.361051 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.361073 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.361081 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.361092 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.361102 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:31Z","lastTransitionTime":"2026-02-18T14:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.463941 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.463985 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.463994 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.464009 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.464018 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:31Z","lastTransitionTime":"2026-02-18T14:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.506867 4896 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.566632 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.566673 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.566684 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.566701 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.566713 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:31Z","lastTransitionTime":"2026-02-18T14:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.669765 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.669800 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.669808 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.669825 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.669833 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:31Z","lastTransitionTime":"2026-02-18T14:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.772161 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.772434 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.772442 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.772456 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.772466 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:31Z","lastTransitionTime":"2026-02-18T14:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.871732 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 07:03:02.571682599 +0000 UTC Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.874285 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.874313 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.874324 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.874339 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.874348 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:31Z","lastTransitionTime":"2026-02-18T14:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.939827 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:31Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.950803 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:31Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.963196 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:31Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.973074 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6wprd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8896bb9-c6fe-4111-a49a-9dadf23e6855\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80d2101411bb762554b416ff237ce9ce2455e43c3c5379118322264ffd9dcd94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5jc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6wprd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:31Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.975989 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.976028 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.976038 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.976053 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.976064 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:31Z","lastTransitionTime":"2026-02-18T14:55:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:31 crc kubenswrapper[4896]: I0218 14:55:31.993587 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:31Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.007529 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.019699 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.031003 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.043110 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.055342 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.068380 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.077763 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.077812 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.077825 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.077842 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.077854 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:32Z","lastTransitionTime":"2026-02-18T14:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.080624 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.092123 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.106196 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.123807 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.156290 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" event={"ID":"449d40af-fe23-4350-b66f-de4ce6614177","Type":"ContainerStarted","Data":"b4edff5db04515901df499d9140f2cf34c48e136b394b098943d8559647894bf"} Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.156530 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.159596 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" event={"ID":"e2efa13d-b712-4f54-967a-d948be1fce0e","Type":"ContainerStarted","Data":"4e6e1900020612d617118e0cca524ffaafe1fac28997fcf1f73a75fe743b4f55"} Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.169075 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.179757 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.179789 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.179800 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.179813 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.179822 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:32Z","lastTransitionTime":"2026-02-18T14:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.181095 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.186836 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.194203 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.208418 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.219216 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.229914 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.245346 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4edff5db04515901df499d9140f2cf34c48e136b394b098943d8559647894bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.255592 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6wprd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8896bb9-c6fe-4111-a49a-9dadf23e6855\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80d2101411bb762554b416ff237ce9ce2455e43c3c5379118322264ffd9dcd94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5jc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6wprd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.266941 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.278767 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.282543 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.282571 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.282583 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.282598 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.282610 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:32Z","lastTransitionTime":"2026-02-18T14:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.291934 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.306007 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.331076 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.348406 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.361799 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.374641 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.386064 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.386124 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.386143 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.386176 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.386199 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:32Z","lastTransitionTime":"2026-02-18T14:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.424386 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4edff5db04515901df499d9140f2cf34c48e136b394b098943d8559647894bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.438966 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6wprd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8896bb9-c6fe-4111-a49a-9dadf23e6855\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80d2101411bb762554b416ff237ce9ce2455e43c3c5379118322264ffd9dcd94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5jc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6wprd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.455341 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.469444 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.486049 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6e1900020612d617118e0cca524ffaafe1fac28997fcf1f73a75fe743b4f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.488826 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.488906 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.488924 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.488952 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.488970 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:32Z","lastTransitionTime":"2026-02-18T14:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.500415 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.518784 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.537759 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.555514 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.569055 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.586977 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.591062 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.591103 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.591115 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.591135 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.591150 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:32Z","lastTransitionTime":"2026-02-18T14:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.604474 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.618356 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.632311 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:32Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.693537 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.693599 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.693612 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.693630 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.693642 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:32Z","lastTransitionTime":"2026-02-18T14:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.796580 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.796652 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.796672 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.796696 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.796714 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:32Z","lastTransitionTime":"2026-02-18T14:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.872598 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 13:18:28.47116667 +0000 UTC Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.899439 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.899480 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.899488 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.899502 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.899511 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:32Z","lastTransitionTime":"2026-02-18T14:55:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.927189 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.927216 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:55:32 crc kubenswrapper[4896]: E0218 14:55:32.927346 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:55:32 crc kubenswrapper[4896]: I0218 14:55:32.927493 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:55:32 crc kubenswrapper[4896]: E0218 14:55:32.927637 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:55:32 crc kubenswrapper[4896]: E0218 14:55:32.927495 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.002565 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.002620 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.002636 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.002689 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.002706 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:33Z","lastTransitionTime":"2026-02-18T14:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.105080 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.105111 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.105121 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.105138 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.105150 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:33Z","lastTransitionTime":"2026-02-18T14:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.161923 4896 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.162719 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.190003 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.203905 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:33Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.207975 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.208005 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.208014 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.208027 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.208036 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:33Z","lastTransitionTime":"2026-02-18T14:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.223499 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4edff5db04515901df499d9140f2cf34c48e136b394b098943d8559647894bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:33Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.246177 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6e1900020612d617118e0cca524ffaafe1fac28997fcf1f73a75fe743b4f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:33Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.257784 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6wprd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8896bb9-c6fe-4111-a49a-9dadf23e6855\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80d2101411bb762554b416ff237ce9ce2455e43c3c5379118322264ffd9dcd94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5jc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6wprd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:33Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.268395 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:33Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.278077 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:33Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.289136 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:33Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.298746 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:33Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.310147 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.310189 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.310200 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.310216 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.310227 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:33Z","lastTransitionTime":"2026-02-18T14:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.316460 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:33Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.328773 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:33Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.341360 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:33Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.350758 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:33Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.361353 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:33Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.372284 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:33Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.384826 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:33Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.412819 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.412858 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.412869 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.412882 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.412890 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:33Z","lastTransitionTime":"2026-02-18T14:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.515113 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.515146 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.515157 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.515170 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.515180 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:33Z","lastTransitionTime":"2026-02-18T14:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.617919 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.618084 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.618093 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.618107 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.618115 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:33Z","lastTransitionTime":"2026-02-18T14:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.771538 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.771791 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.771876 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.771958 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.772033 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:33Z","lastTransitionTime":"2026-02-18T14:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.876889 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 07:41:16.596418514 +0000 UTC Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.880706 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.880896 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.881080 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.881196 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.881318 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:33Z","lastTransitionTime":"2026-02-18T14:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.983664 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.984175 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.984187 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.984210 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:33 crc kubenswrapper[4896]: I0218 14:55:33.984221 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:33Z","lastTransitionTime":"2026-02-18T14:55:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.086126 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.086161 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.086170 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.086181 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.086191 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:34Z","lastTransitionTime":"2026-02-18T14:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.165521 4896 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.188530 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.188562 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.188572 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.188587 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.188597 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:34Z","lastTransitionTime":"2026-02-18T14:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.291367 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.291422 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.291430 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.291444 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.291468 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:34Z","lastTransitionTime":"2026-02-18T14:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.394810 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.394877 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.394889 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.394907 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.394923 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:34Z","lastTransitionTime":"2026-02-18T14:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.497369 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.497400 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.497409 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.497422 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.497431 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:34Z","lastTransitionTime":"2026-02-18T14:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.603389 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.603425 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.603434 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.603447 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.603456 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:34Z","lastTransitionTime":"2026-02-18T14:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.705750 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.705783 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.705792 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.705804 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.705812 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:34Z","lastTransitionTime":"2026-02-18T14:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.808211 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.808268 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.808276 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.808289 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.808299 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:34Z","lastTransitionTime":"2026-02-18T14:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.877365 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 19:17:31.255797712 +0000 UTC Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.910180 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.910224 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.910260 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.910333 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.910349 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:34Z","lastTransitionTime":"2026-02-18T14:55:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.927670 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.927688 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:55:34 crc kubenswrapper[4896]: I0218 14:55:34.927697 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:55:34 crc kubenswrapper[4896]: E0218 14:55:34.927755 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:55:34 crc kubenswrapper[4896]: E0218 14:55:34.927820 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:55:34 crc kubenswrapper[4896]: E0218 14:55:34.927892 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.012530 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.012561 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.012586 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.012600 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.012609 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:35Z","lastTransitionTime":"2026-02-18T14:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.114931 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.114962 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.114970 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.114982 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.114990 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:35Z","lastTransitionTime":"2026-02-18T14:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.168037 4896 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.217591 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.217649 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.217660 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.217679 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.217692 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:35Z","lastTransitionTime":"2026-02-18T14:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.319686 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.320448 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.320542 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.320616 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.320684 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:35Z","lastTransitionTime":"2026-02-18T14:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.423264 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.423312 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.423326 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.423347 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.423360 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:35Z","lastTransitionTime":"2026-02-18T14:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.525753 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.525791 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.525799 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.525813 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.525822 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:35Z","lastTransitionTime":"2026-02-18T14:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.627703 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.627743 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.627755 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.627771 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.627783 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:35Z","lastTransitionTime":"2026-02-18T14:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.730226 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.730285 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.730296 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.730314 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.730326 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:35Z","lastTransitionTime":"2026-02-18T14:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.832068 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.832101 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.832111 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.832140 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.832150 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:35Z","lastTransitionTime":"2026-02-18T14:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.878084 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 19:40:21.106703896 +0000 UTC Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.934804 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.934835 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.934843 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.934865 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:35 crc kubenswrapper[4896]: I0218 14:55:35.934875 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:35Z","lastTransitionTime":"2026-02-18T14:55:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.039465 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.039526 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.039539 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.039565 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.039587 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:36Z","lastTransitionTime":"2026-02-18T14:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.141905 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.141979 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.142000 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.142031 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.142052 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:36Z","lastTransitionTime":"2026-02-18T14:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.173975 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w862c_449d40af-fe23-4350-b66f-de4ce6614177/ovnkube-controller/0.log" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.177586 4896 generic.go:334] "Generic (PLEG): container finished" podID="449d40af-fe23-4350-b66f-de4ce6614177" containerID="b4edff5db04515901df499d9140f2cf34c48e136b394b098943d8559647894bf" exitCode=1 Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.177643 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" event={"ID":"449d40af-fe23-4350-b66f-de4ce6614177","Type":"ContainerDied","Data":"b4edff5db04515901df499d9140f2cf34c48e136b394b098943d8559647894bf"} Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.178551 4896 scope.go:117] "RemoveContainer" containerID="b4edff5db04515901df499d9140f2cf34c48e136b394b098943d8559647894bf" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.197897 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:36Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.221858 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:36Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.236310 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:36Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.245122 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.245569 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.245651 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.245737 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.245814 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:36Z","lastTransitionTime":"2026-02-18T14:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.251834 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:36Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.267158 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:36Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.282017 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:36Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.302603 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b4edff5db04515901df499d9140f2cf34c48e136b394b098943d8559647894bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4edff5db04515901df499d9140f2cf34c48e136b394b098943d8559647894bf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:55:35Z\\\",\\\"message\\\":\\\"5.925083 6210 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 14:55:35.925127 6210 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0218 14:55:35.925129 6210 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 14:55:35.925361 6210 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 14:55:35.925534 6210 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0218 14:55:35.925557 6210 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0218 14:55:35.925576 6210 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0218 14:55:35.925673 6210 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0218 14:55:35.925739 6210 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:36Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.314855 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6wprd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8896bb9-c6fe-4111-a49a-9dadf23e6855\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80d2101411bb762554b416ff237ce9ce2455e43c3c5379118322264ffd9dcd94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5jc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6wprd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:36Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.332073 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:36Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.344451 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:36Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.349168 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.349205 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.349216 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.349244 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.349255 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:36Z","lastTransitionTime":"2026-02-18T14:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.360677 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6e1900020612d617118e0cca524ffaafe1fac28997fcf1f73a75fe743b4f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:36Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.372739 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:36Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.393135 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:36Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.413795 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:36Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.433348 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:36Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.453000 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.453030 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.453042 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.453061 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.453072 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:36Z","lastTransitionTime":"2026-02-18T14:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.555764 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.555804 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.555816 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.556219 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.556272 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:36Z","lastTransitionTime":"2026-02-18T14:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.607608 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.658428 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.658455 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.658462 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.658475 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.658484 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:36Z","lastTransitionTime":"2026-02-18T14:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.761855 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.761911 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.761930 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.761950 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.761960 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:36Z","lastTransitionTime":"2026-02-18T14:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.865505 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.865577 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.865595 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.865622 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.865643 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:36Z","lastTransitionTime":"2026-02-18T14:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.878737 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 17:28:50.87737784 +0000 UTC Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.927297 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.927394 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:55:36 crc kubenswrapper[4896]: I0218 14:55:36.927465 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:55:36 crc kubenswrapper[4896]: E0218 14:55:36.927537 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:55:36 crc kubenswrapper[4896]: E0218 14:55:36.927717 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:55:36 crc kubenswrapper[4896]: E0218 14:55:36.928000 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:36.968505 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:36.968528 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:36.968536 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:36.968548 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:36.968556 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:36Z","lastTransitionTime":"2026-02-18T14:55:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.070915 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.070959 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.070972 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.070991 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.071001 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:37Z","lastTransitionTime":"2026-02-18T14:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.173479 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.173520 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.173530 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.173546 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.173555 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:37Z","lastTransitionTime":"2026-02-18T14:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.181430 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w862c_449d40af-fe23-4350-b66f-de4ce6614177/ovnkube-controller/0.log" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.184436 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" event={"ID":"449d40af-fe23-4350-b66f-de4ce6614177","Type":"ContainerStarted","Data":"54aecfeba56baae4b078096062414aca0ae6e1769fbbc971a15b1e84bfac0538"} Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.184705 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.200661 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:37Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.219353 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:37Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.241142 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6e1900020612d617118e0cca524ffaafe1fac28997fcf1f73a75fe743b4f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:37Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.254450 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6wprd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8896bb9-c6fe-4111-a49a-9dadf23e6855\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80d2101411bb762554b416ff237ce9ce2455e43c3c5379118322264ffd9dcd94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5jc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6wprd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:37Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.273438 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:37Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.275197 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.275256 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.275269 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.275286 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.275298 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:37Z","lastTransitionTime":"2026-02-18T14:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.289877 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:37Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.309666 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:37Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.325010 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:37Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.340759 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:37Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.357853 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:37Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.375888 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:37Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.377455 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.377503 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.377516 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.377538 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.377555 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:37Z","lastTransitionTime":"2026-02-18T14:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.400045 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:37Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.435786 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:37Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.457694 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:37Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.478888 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54aecfeba56baae4b078096062414aca0ae6e1769fbbc971a15b1e84bfac0538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4edff5db04515901df499d9140f2cf34c48e136b394b098943d8559647894bf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:55:35Z\\\",\\\"message\\\":\\\"5.925083 6210 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 14:55:35.925127 6210 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0218 14:55:35.925129 6210 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 14:55:35.925361 6210 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 14:55:35.925534 6210 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0218 14:55:35.925557 6210 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0218 14:55:35.925576 6210 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0218 14:55:35.925673 6210 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0218 14:55:35.925739 6210 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:37Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.479752 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.479796 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.479813 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.479832 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.479844 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:37Z","lastTransitionTime":"2026-02-18T14:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.573114 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb"] Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.573694 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.575633 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.576949 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.581722 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.581769 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.581780 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.581799 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.581810 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:37Z","lastTransitionTime":"2026-02-18T14:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.596844 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:37Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.611127 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:37Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.625599 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:37Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.638545 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:37Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.648326 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:37Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.661345 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32bd695f-cf0b-425a-b764-87b3f678d4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8j9gb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:37Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.672885 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:37Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.684020 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.684051 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.684059 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.684072 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.684081 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:37Z","lastTransitionTime":"2026-02-18T14:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.689546 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54aecfeba56baae4b078096062414aca0ae6e1769fbbc971a15b1e84bfac0538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4edff5db04515901df499d9140f2cf34c48e136b394b098943d8559647894bf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:55:35Z\\\",\\\"message\\\":\\\"5.925083 6210 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 14:55:35.925127 6210 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0218 14:55:35.925129 6210 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 14:55:35.925361 6210 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 14:55:35.925534 6210 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0218 14:55:35.925557 6210 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0218 14:55:35.925576 6210 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0218 14:55:35.925673 6210 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0218 14:55:35.925739 6210 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:37Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.699777 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:37Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.709066 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:37Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.720929 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6e1900020612d617118e0cca524ffaafe1fac28997fcf1f73a75fe743b4f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:37Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.729642 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6wprd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8896bb9-c6fe-4111-a49a-9dadf23e6855\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80d2101411bb762554b416ff237ce9ce2455e43c3c5379118322264ffd9dcd94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5jc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6wprd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:37Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.746080 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:37Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.757146 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:37Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.767915 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:37Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.770223 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jk59\" (UniqueName: \"kubernetes.io/projected/32bd695f-cf0b-425a-b764-87b3f678d4ea-kube-api-access-5jk59\") pod \"ovnkube-control-plane-749d76644c-8j9gb\" (UID: \"32bd695f-cf0b-425a-b764-87b3f678d4ea\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.770382 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/32bd695f-cf0b-425a-b764-87b3f678d4ea-env-overrides\") pod \"ovnkube-control-plane-749d76644c-8j9gb\" (UID: \"32bd695f-cf0b-425a-b764-87b3f678d4ea\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.770464 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/32bd695f-cf0b-425a-b764-87b3f678d4ea-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-8j9gb\" (UID: \"32bd695f-cf0b-425a-b764-87b3f678d4ea\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.770628 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/32bd695f-cf0b-425a-b764-87b3f678d4ea-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-8j9gb\" (UID: \"32bd695f-cf0b-425a-b764-87b3f678d4ea\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.777386 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:37Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.785720 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.785757 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.785767 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.785783 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.785794 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:37Z","lastTransitionTime":"2026-02-18T14:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.871466 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jk59\" (UniqueName: \"kubernetes.io/projected/32bd695f-cf0b-425a-b764-87b3f678d4ea-kube-api-access-5jk59\") pod \"ovnkube-control-plane-749d76644c-8j9gb\" (UID: \"32bd695f-cf0b-425a-b764-87b3f678d4ea\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.871512 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/32bd695f-cf0b-425a-b764-87b3f678d4ea-env-overrides\") pod \"ovnkube-control-plane-749d76644c-8j9gb\" (UID: \"32bd695f-cf0b-425a-b764-87b3f678d4ea\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.871537 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/32bd695f-cf0b-425a-b764-87b3f678d4ea-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-8j9gb\" (UID: \"32bd695f-cf0b-425a-b764-87b3f678d4ea\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.871573 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/32bd695f-cf0b-425a-b764-87b3f678d4ea-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-8j9gb\" (UID: \"32bd695f-cf0b-425a-b764-87b3f678d4ea\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.872328 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/32bd695f-cf0b-425a-b764-87b3f678d4ea-env-overrides\") pod \"ovnkube-control-plane-749d76644c-8j9gb\" (UID: \"32bd695f-cf0b-425a-b764-87b3f678d4ea\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.872406 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/32bd695f-cf0b-425a-b764-87b3f678d4ea-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-8j9gb\" (UID: \"32bd695f-cf0b-425a-b764-87b3f678d4ea\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.879359 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 06:32:01.620377556 +0000 UTC Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.879951 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/32bd695f-cf0b-425a-b764-87b3f678d4ea-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-8j9gb\" (UID: \"32bd695f-cf0b-425a-b764-87b3f678d4ea\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.887294 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.887467 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.887566 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.887654 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.887734 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:37Z","lastTransitionTime":"2026-02-18T14:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.888468 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jk59\" (UniqueName: \"kubernetes.io/projected/32bd695f-cf0b-425a-b764-87b3f678d4ea-kube-api-access-5jk59\") pod \"ovnkube-control-plane-749d76644c-8j9gb\" (UID: \"32bd695f-cf0b-425a-b764-87b3f678d4ea\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.990469 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.990506 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.990516 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.990532 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:37 crc kubenswrapper[4896]: I0218 14:55:37.990542 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:37Z","lastTransitionTime":"2026-02-18T14:55:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.092205 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.092307 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.092321 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.092337 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.092348 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:38Z","lastTransitionTime":"2026-02-18T14:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.188086 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.188985 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w862c_449d40af-fe23-4350-b66f-de4ce6614177/ovnkube-controller/1.log" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.189515 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w862c_449d40af-fe23-4350-b66f-de4ce6614177/ovnkube-controller/0.log" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.192828 4896 generic.go:334] "Generic (PLEG): container finished" podID="449d40af-fe23-4350-b66f-de4ce6614177" containerID="54aecfeba56baae4b078096062414aca0ae6e1769fbbc971a15b1e84bfac0538" exitCode=1 Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.192875 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" event={"ID":"449d40af-fe23-4350-b66f-de4ce6614177","Type":"ContainerDied","Data":"54aecfeba56baae4b078096062414aca0ae6e1769fbbc971a15b1e84bfac0538"} Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.192939 4896 scope.go:117] "RemoveContainer" containerID="b4edff5db04515901df499d9140f2cf34c48e136b394b098943d8559647894bf" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.193360 4896 scope.go:117] "RemoveContainer" containerID="54aecfeba56baae4b078096062414aca0ae6e1769fbbc971a15b1e84bfac0538" Feb 18 14:55:38 crc kubenswrapper[4896]: E0218 14:55:38.193505 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-w862c_openshift-ovn-kubernetes(449d40af-fe23-4350-b66f-de4ce6614177)\"" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" podUID="449d40af-fe23-4350-b66f-de4ce6614177" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.193950 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.193978 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.193986 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.193999 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.194008 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:38Z","lastTransitionTime":"2026-02-18T14:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.225386 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.241999 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.255363 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.268203 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.279494 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.291229 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.298205 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-98x74"] Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.298952 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:55:38 crc kubenswrapper[4896]: E0218 14:55:38.299001 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.304604 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.305282 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.305317 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.305330 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.305345 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.305355 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:38Z","lastTransitionTime":"2026-02-18T14:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.315349 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.325284 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.335455 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32bd695f-cf0b-425a-b764-87b3f678d4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8j9gb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.346497 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.364581 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54aecfeba56baae4b078096062414aca0ae6e1769fbbc971a15b1e84bfac0538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4edff5db04515901df499d9140f2cf34c48e136b394b098943d8559647894bf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:55:35Z\\\",\\\"message\\\":\\\"5.925083 6210 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 14:55:35.925127 6210 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0218 14:55:35.925129 6210 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 14:55:35.925361 6210 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 14:55:35.925534 6210 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0218 14:55:35.925557 6210 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0218 14:55:35.925576 6210 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0218 14:55:35.925673 6210 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0218 14:55:35.925739 6210 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54aecfeba56baae4b078096062414aca0ae6e1769fbbc971a15b1e84bfac0538\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"message\\\":\\\"umn _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0218 14:55:37.429578 6353 ovnkube.go:599] Stopped ovnkube\\\\nI0218 14:55:37.429657 6353 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0218 14:55:37.429778 6353 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.374456 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.384025 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.395905 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6e1900020612d617118e0cca524ffaafe1fac28997fcf1f73a75fe743b4f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.406753 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6wprd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8896bb9-c6fe-4111-a49a-9dadf23e6855\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80d2101411bb762554b416ff237ce9ce2455e43c3c5379118322264ffd9dcd94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5jc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6wprd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.407252 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.407284 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.407299 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.407312 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.407321 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:38Z","lastTransitionTime":"2026-02-18T14:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.420071 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.433453 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.444918 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.456922 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32bd695f-cf0b-425a-b764-87b3f678d4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8j9gb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.466926 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-98x74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"229213f7-7a25-4982-a449-0122912ba05c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-98x74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.478546 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5rmp\" (UniqueName: \"kubernetes.io/projected/229213f7-7a25-4982-a449-0122912ba05c-kube-api-access-h5rmp\") pod \"network-metrics-daemon-98x74\" (UID: \"229213f7-7a25-4982-a449-0122912ba05c\") " pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.478605 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/229213f7-7a25-4982-a449-0122912ba05c-metrics-certs\") pod \"network-metrics-daemon-98x74\" (UID: \"229213f7-7a25-4982-a449-0122912ba05c\") " pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.478472 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.490855 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.502486 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.510575 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.510603 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.510626 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.510639 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.510653 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:38Z","lastTransitionTime":"2026-02-18T14:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.519076 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54aecfeba56baae4b078096062414aca0ae6e1769fbbc971a15b1e84bfac0538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4edff5db04515901df499d9140f2cf34c48e136b394b098943d8559647894bf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:55:35Z\\\",\\\"message\\\":\\\"5.925083 6210 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 14:55:35.925127 6210 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0218 14:55:35.925129 6210 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 14:55:35.925361 6210 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 14:55:35.925534 6210 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0218 14:55:35.925557 6210 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0218 14:55:35.925576 6210 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0218 14:55:35.925673 6210 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0218 14:55:35.925739 6210 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54aecfeba56baae4b078096062414aca0ae6e1769fbbc971a15b1e84bfac0538\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"message\\\":\\\"umn _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0218 14:55:37.429578 6353 ovnkube.go:599] Stopped ovnkube\\\\nI0218 14:55:37.429657 6353 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0218 14:55:37.429778 6353 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.532910 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.547745 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6e1900020612d617118e0cca524ffaafe1fac28997fcf1f73a75fe743b4f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.561828 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6wprd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8896bb9-c6fe-4111-a49a-9dadf23e6855\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80d2101411bb762554b416ff237ce9ce2455e43c3c5379118322264ffd9dcd94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5jc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6wprd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.577275 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.579691 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/229213f7-7a25-4982-a449-0122912ba05c-metrics-certs\") pod \"network-metrics-daemon-98x74\" (UID: \"229213f7-7a25-4982-a449-0122912ba05c\") " pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.579797 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5rmp\" (UniqueName: \"kubernetes.io/projected/229213f7-7a25-4982-a449-0122912ba05c-kube-api-access-h5rmp\") pod \"network-metrics-daemon-98x74\" (UID: \"229213f7-7a25-4982-a449-0122912ba05c\") " pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:55:38 crc kubenswrapper[4896]: E0218 14:55:38.579902 4896 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 14:55:38 crc kubenswrapper[4896]: E0218 14:55:38.579963 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/229213f7-7a25-4982-a449-0122912ba05c-metrics-certs podName:229213f7-7a25-4982-a449-0122912ba05c nodeName:}" failed. No retries permitted until 2026-02-18 14:55:39.079944368 +0000 UTC m=+38.009278581 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/229213f7-7a25-4982-a449-0122912ba05c-metrics-certs") pod "network-metrics-daemon-98x74" (UID: "229213f7-7a25-4982-a449-0122912ba05c") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.591844 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.595309 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5rmp\" (UniqueName: \"kubernetes.io/projected/229213f7-7a25-4982-a449-0122912ba05c-kube-api-access-h5rmp\") pod \"network-metrics-daemon-98x74\" (UID: \"229213f7-7a25-4982-a449-0122912ba05c\") " pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.603607 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.612386 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.612409 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.612417 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.612431 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.612440 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:38Z","lastTransitionTime":"2026-02-18T14:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.621004 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.632368 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.680830 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.680926 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.680957 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.680986 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.681002 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:55:38 crc kubenswrapper[4896]: E0218 14:55:38.683552 4896 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 14:55:38 crc kubenswrapper[4896]: E0218 14:55:38.683585 4896 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 14:55:38 crc kubenswrapper[4896]: E0218 14:55:38.683622 4896 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 14:55:38 crc kubenswrapper[4896]: E0218 14:55:38.683640 4896 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 14:55:38 crc kubenswrapper[4896]: E0218 14:55:38.683638 4896 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 14:55:38 crc kubenswrapper[4896]: E0218 14:55:38.683591 4896 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 14:55:38 crc kubenswrapper[4896]: E0218 14:55:38.683728 4896 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 14:55:38 crc kubenswrapper[4896]: E0218 14:55:38.683558 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:55:54.683517579 +0000 UTC m=+53.612851792 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:55:38 crc kubenswrapper[4896]: E0218 14:55:38.683780 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-18 14:55:54.683769046 +0000 UTC m=+53.613103259 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 14:55:38 crc kubenswrapper[4896]: E0218 14:55:38.683796 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 14:55:54.683787636 +0000 UTC m=+53.613121839 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 14:55:38 crc kubenswrapper[4896]: E0218 14:55:38.683814 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-18 14:55:54.683801586 +0000 UTC m=+53.613135799 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 14:55:38 crc kubenswrapper[4896]: E0218 14:55:38.685350 4896 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 14:55:38 crc kubenswrapper[4896]: E0218 14:55:38.685487 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 14:55:54.685410889 +0000 UTC m=+53.614745102 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.714938 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.714995 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.715005 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.715023 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.715034 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:38Z","lastTransitionTime":"2026-02-18T14:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.817685 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.817771 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.817784 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.817809 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.817825 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:38Z","lastTransitionTime":"2026-02-18T14:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.880203 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 09:16:09.335371524 +0000 UTC Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.905370 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.905424 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.905437 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.905460 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.905473 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:38Z","lastTransitionTime":"2026-02-18T14:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:38 crc kubenswrapper[4896]: E0218 14:55:38.917743 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3d8f65ef-15b5-4cc3-be66-c8dd6cb956e9\\\",\\\"systemUUID\\\":\\\"d255e81f-4350-4697-bbdf-3f95dec57eda\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.921119 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.921158 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.921170 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.921189 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.921199 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:38Z","lastTransitionTime":"2026-02-18T14:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.927633 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:55:38 crc kubenswrapper[4896]: E0218 14:55:38.927745 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.927849 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:55:38 crc kubenswrapper[4896]: E0218 14:55:38.927966 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.928057 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:55:38 crc kubenswrapper[4896]: E0218 14:55:38.928128 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:55:38 crc kubenswrapper[4896]: E0218 14:55:38.931479 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3d8f65ef-15b5-4cc3-be66-c8dd6cb956e9\\\",\\\"systemUUID\\\":\\\"d255e81f-4350-4697-bbdf-3f95dec57eda\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.935324 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.935362 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.935372 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.935390 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.935399 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:38Z","lastTransitionTime":"2026-02-18T14:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:38 crc kubenswrapper[4896]: E0218 14:55:38.946154 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3d8f65ef-15b5-4cc3-be66-c8dd6cb956e9\\\",\\\"systemUUID\\\":\\\"d255e81f-4350-4697-bbdf-3f95dec57eda\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.949165 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.949200 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.949209 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.949224 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.949250 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:38Z","lastTransitionTime":"2026-02-18T14:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:38 crc kubenswrapper[4896]: E0218 14:55:38.959887 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3d8f65ef-15b5-4cc3-be66-c8dd6cb956e9\\\",\\\"systemUUID\\\":\\\"d255e81f-4350-4697-bbdf-3f95dec57eda\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.962703 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.962748 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.962757 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.962772 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.962781 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:38Z","lastTransitionTime":"2026-02-18T14:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:38 crc kubenswrapper[4896]: E0218 14:55:38.973312 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3d8f65ef-15b5-4cc3-be66-c8dd6cb956e9\\\",\\\"systemUUID\\\":\\\"d255e81f-4350-4697-bbdf-3f95dec57eda\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:38Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:38 crc kubenswrapper[4896]: E0218 14:55:38.973447 4896 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.974765 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.974818 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.974828 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.974843 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:38 crc kubenswrapper[4896]: I0218 14:55:38.974853 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:38Z","lastTransitionTime":"2026-02-18T14:55:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.077525 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.077559 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.077570 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.077585 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.077595 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:39Z","lastTransitionTime":"2026-02-18T14:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.083738 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/229213f7-7a25-4982-a449-0122912ba05c-metrics-certs\") pod \"network-metrics-daemon-98x74\" (UID: \"229213f7-7a25-4982-a449-0122912ba05c\") " pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:55:39 crc kubenswrapper[4896]: E0218 14:55:39.083870 4896 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 14:55:39 crc kubenswrapper[4896]: E0218 14:55:39.083916 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/229213f7-7a25-4982-a449-0122912ba05c-metrics-certs podName:229213f7-7a25-4982-a449-0122912ba05c nodeName:}" failed. No retries permitted until 2026-02-18 14:55:40.083902825 +0000 UTC m=+39.013237038 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/229213f7-7a25-4982-a449-0122912ba05c-metrics-certs") pod "network-metrics-daemon-98x74" (UID: "229213f7-7a25-4982-a449-0122912ba05c") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.179892 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.179931 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.179939 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.179952 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.179962 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:39Z","lastTransitionTime":"2026-02-18T14:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.198315 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" event={"ID":"32bd695f-cf0b-425a-b764-87b3f678d4ea","Type":"ContainerStarted","Data":"f7021f783fefb51e5fb6b701e58eb9279a0513b9ff37cba435d69e1b7bb1d1a8"} Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.198389 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" event={"ID":"32bd695f-cf0b-425a-b764-87b3f678d4ea","Type":"ContainerStarted","Data":"390187c9005f214d18d43cd6b4f987ccc7d2197f8629997ca4c5cd3ffb49c777"} Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.198439 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" event={"ID":"32bd695f-cf0b-425a-b764-87b3f678d4ea","Type":"ContainerStarted","Data":"9a7022ab1c86c6751c7a4ab20ff3c6a86f49d19ae4d2741ebe2d05fea18989a7"} Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.201299 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w862c_449d40af-fe23-4350-b66f-de4ce6614177/ovnkube-controller/1.log" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.205260 4896 scope.go:117] "RemoveContainer" containerID="54aecfeba56baae4b078096062414aca0ae6e1769fbbc971a15b1e84bfac0538" Feb 18 14:55:39 crc kubenswrapper[4896]: E0218 14:55:39.205402 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-w862c_openshift-ovn-kubernetes(449d40af-fe23-4350-b66f-de4ce6614177)\"" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" podUID="449d40af-fe23-4350-b66f-de4ce6614177" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.220041 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.239317 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54aecfeba56baae4b078096062414aca0ae6e1769fbbc971a15b1e84bfac0538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b4edff5db04515901df499d9140f2cf34c48e136b394b098943d8559647894bf\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:55:35Z\\\",\\\"message\\\":\\\"5.925083 6210 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 14:55:35.925127 6210 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0218 14:55:35.925129 6210 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 14:55:35.925361 6210 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0218 14:55:35.925534 6210 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0218 14:55:35.925557 6210 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0218 14:55:35.925576 6210 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0218 14:55:35.925673 6210 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0218 14:55:35.925739 6210 handler.go:208] Removed *v1.EgressIP ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54aecfeba56baae4b078096062414aca0ae6e1769fbbc971a15b1e84bfac0538\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"message\\\":\\\"umn _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0218 14:55:37.429578 6353 ovnkube.go:599] Stopped ovnkube\\\\nI0218 14:55:37.429657 6353 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0218 14:55:37.429778 6353 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.251701 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.267346 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6e1900020612d617118e0cca524ffaafe1fac28997fcf1f73a75fe743b4f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.282269 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.282311 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.282323 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.282342 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.282355 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:39Z","lastTransitionTime":"2026-02-18T14:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.286185 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6wprd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8896bb9-c6fe-4111-a49a-9dadf23e6855\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80d2101411bb762554b416ff237ce9ce2455e43c3c5379118322264ffd9dcd94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5jc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6wprd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.303501 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.317841 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.334787 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.358571 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.378593 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.384564 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.384598 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.384627 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.384646 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.384660 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:39Z","lastTransitionTime":"2026-02-18T14:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.393609 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.404545 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.415619 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.427432 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32bd695f-cf0b-425a-b764-87b3f678d4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://390187c9005f214d18d43cd6b4f987ccc7d2197f8629997ca4c5cd3ffb49c777\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7021f783fefb51e5fb6b701e58eb9279a0513b9ff37cba435d69e1b7bb1d1a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8j9gb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.438874 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-98x74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"229213f7-7a25-4982-a449-0122912ba05c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-98x74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.449937 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.462341 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.473386 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.487254 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.487291 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.487304 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.487320 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.487331 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:39Z","lastTransitionTime":"2026-02-18T14:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.490112 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54aecfeba56baae4b078096062414aca0ae6e1769fbbc971a15b1e84bfac0538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54aecfeba56baae4b078096062414aca0ae6e1769fbbc971a15b1e84bfac0538\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"message\\\":\\\"umn _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0218 14:55:37.429578 6353 ovnkube.go:599] Stopped ovnkube\\\\nI0218 14:55:37.429657 6353 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0218 14:55:37.429778 6353 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-w862c_openshift-ovn-kubernetes(449d40af-fe23-4350-b66f-de4ce6614177)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.500576 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.512372 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6e1900020612d617118e0cca524ffaafe1fac28997fcf1f73a75fe743b4f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.520047 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6wprd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8896bb9-c6fe-4111-a49a-9dadf23e6855\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80d2101411bb762554b416ff237ce9ce2455e43c3c5379118322264ffd9dcd94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5jc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6wprd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.529440 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.539373 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.547317 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.564324 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.577929 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.588904 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.588934 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.588943 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.588956 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.588964 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:39Z","lastTransitionTime":"2026-02-18T14:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.590373 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.602318 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.611025 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.623903 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32bd695f-cf0b-425a-b764-87b3f678d4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://390187c9005f214d18d43cd6b4f987ccc7d2197f8629997ca4c5cd3ffb49c777\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7021f783fefb51e5fb6b701e58eb9279a0513b9ff37cba435d69e1b7bb1d1a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8j9gb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.632459 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-98x74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"229213f7-7a25-4982-a449-0122912ba05c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-98x74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.642323 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.652339 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:39Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.691153 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.691191 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.691198 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.691212 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.691221 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:39Z","lastTransitionTime":"2026-02-18T14:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.793446 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.793476 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.793484 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.793497 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.793507 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:39Z","lastTransitionTime":"2026-02-18T14:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.880328 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 18:41:39.699314709 +0000 UTC Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.895343 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.895400 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.895423 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.895449 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.895472 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:39Z","lastTransitionTime":"2026-02-18T14:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.927918 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:55:39 crc kubenswrapper[4896]: E0218 14:55:39.928109 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.997654 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.997905 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.997979 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.998060 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:39 crc kubenswrapper[4896]: I0218 14:55:39.998366 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:39Z","lastTransitionTime":"2026-02-18T14:55:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.091152 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/229213f7-7a25-4982-a449-0122912ba05c-metrics-certs\") pod \"network-metrics-daemon-98x74\" (UID: \"229213f7-7a25-4982-a449-0122912ba05c\") " pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:55:40 crc kubenswrapper[4896]: E0218 14:55:40.091349 4896 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 14:55:40 crc kubenswrapper[4896]: E0218 14:55:40.091439 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/229213f7-7a25-4982-a449-0122912ba05c-metrics-certs podName:229213f7-7a25-4982-a449-0122912ba05c nodeName:}" failed. No retries permitted until 2026-02-18 14:55:42.091421787 +0000 UTC m=+41.020756000 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/229213f7-7a25-4982-a449-0122912ba05c-metrics-certs") pod "network-metrics-daemon-98x74" (UID: "229213f7-7a25-4982-a449-0122912ba05c") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.100930 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.101151 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.101261 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.101337 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.101414 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:40Z","lastTransitionTime":"2026-02-18T14:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.203493 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.203534 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.203544 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.203558 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.203569 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:40Z","lastTransitionTime":"2026-02-18T14:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.306098 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.306136 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.306146 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.306162 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.306173 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:40Z","lastTransitionTime":"2026-02-18T14:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.408302 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.408332 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.408341 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.408353 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.408362 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:40Z","lastTransitionTime":"2026-02-18T14:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.511087 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.511144 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.511162 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.511186 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.511206 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:40Z","lastTransitionTime":"2026-02-18T14:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.614657 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.614709 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.614722 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.614739 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.614752 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:40Z","lastTransitionTime":"2026-02-18T14:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.718364 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.718443 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.718468 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.718509 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.718537 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:40Z","lastTransitionTime":"2026-02-18T14:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.821529 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.821603 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.821618 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.821637 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.821654 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:40Z","lastTransitionTime":"2026-02-18T14:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.881311 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 17:20:30.943918597 +0000 UTC Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.923994 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.924062 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.924081 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.924105 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.924122 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:40Z","lastTransitionTime":"2026-02-18T14:55:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.927316 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.927515 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:55:40 crc kubenswrapper[4896]: E0218 14:55:40.927575 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.927598 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:55:40 crc kubenswrapper[4896]: E0218 14:55:40.927741 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:55:40 crc kubenswrapper[4896]: E0218 14:55:40.927925 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:55:40 crc kubenswrapper[4896]: I0218 14:55:40.928467 4896 scope.go:117] "RemoveContainer" containerID="fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.027491 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.027885 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.027898 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.027912 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.027922 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:41Z","lastTransitionTime":"2026-02-18T14:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.129890 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.129930 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.129938 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.129955 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.129996 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:41Z","lastTransitionTime":"2026-02-18T14:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.215223 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.218535 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"646e1cf96a28a07565ec2d56a4ffc32c10e95356b7ebf2399b6b3ba833274a68"} Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.219025 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.233007 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.233056 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.233069 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.233088 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.233101 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:41Z","lastTransitionTime":"2026-02-18T14:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.242294 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54aecfeba56baae4b078096062414aca0ae6e1769fbbc971a15b1e84bfac0538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54aecfeba56baae4b078096062414aca0ae6e1769fbbc971a15b1e84bfac0538\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"message\\\":\\\"umn _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0218 14:55:37.429578 6353 ovnkube.go:599] Stopped ovnkube\\\\nI0218 14:55:37.429657 6353 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0218 14:55:37.429778 6353 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-w862c_openshift-ovn-kubernetes(449d40af-fe23-4350-b66f-de4ce6614177)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:41Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.259272 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:41Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.276022 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:41Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.293006 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:41Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.310732 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6e1900020612d617118e0cca524ffaafe1fac28997fcf1f73a75fe743b4f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:41Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.324826 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6wprd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8896bb9-c6fe-4111-a49a-9dadf23e6855\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80d2101411bb762554b416ff237ce9ce2455e43c3c5379118322264ffd9dcd94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5jc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6wprd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:41Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.336218 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.336314 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.336333 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.336365 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.336386 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:41Z","lastTransitionTime":"2026-02-18T14:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.341871 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e1cf96a28a07565ec2d56a4ffc32c10e95356b7ebf2399b6b3ba833274a68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:41Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.359340 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:41Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.374957 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:41Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.401890 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:41Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.414279 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:41Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.428311 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:41Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.439263 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.439315 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.439328 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.439348 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.439362 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:41Z","lastTransitionTime":"2026-02-18T14:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.441998 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:41Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.454135 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:41Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.467731 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32bd695f-cf0b-425a-b764-87b3f678d4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://390187c9005f214d18d43cd6b4f987ccc7d2197f8629997ca4c5cd3ffb49c777\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7021f783fefb51e5fb6b701e58eb9279a0513b9ff37cba435d69e1b7bb1d1a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8j9gb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:41Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.484723 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-98x74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"229213f7-7a25-4982-a449-0122912ba05c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-98x74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:41Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.509557 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:41Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.542107 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.542166 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.542179 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.542200 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.542214 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:41Z","lastTransitionTime":"2026-02-18T14:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.645981 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.646067 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.646094 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.646132 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.646159 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:41Z","lastTransitionTime":"2026-02-18T14:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.749289 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.749339 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.749349 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.749366 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.749383 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:41Z","lastTransitionTime":"2026-02-18T14:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.852188 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.852322 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.852341 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.852371 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.852392 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:41Z","lastTransitionTime":"2026-02-18T14:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.882103 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 01:38:56.072908485 +0000 UTC Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.927640 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:55:41 crc kubenswrapper[4896]: E0218 14:55:41.927896 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.944520 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:41Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.955026 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.955064 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.955078 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.955095 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.955107 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:41Z","lastTransitionTime":"2026-02-18T14:55:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.956938 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:41Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.975341 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:41Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:41 crc kubenswrapper[4896]: I0218 14:55:41.989489 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e1cf96a28a07565ec2d56a4ffc32c10e95356b7ebf2399b6b3ba833274a68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:41Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.000787 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:41Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.031625 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:42Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.047127 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:42Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.056455 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.056509 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.056525 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.056548 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.056567 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:42Z","lastTransitionTime":"2026-02-18T14:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.070257 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32bd695f-cf0b-425a-b764-87b3f678d4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://390187c9005f214d18d43cd6b4f987ccc7d2197f8629997ca4c5cd3ffb49c777\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7021f783fefb51e5fb6b701e58eb9279a0513b9ff37cba435d69e1b7bb1d1a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8j9gb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:42Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.084741 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-98x74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"229213f7-7a25-4982-a449-0122912ba05c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-98x74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:42Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.104794 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:42Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.114182 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/229213f7-7a25-4982-a449-0122912ba05c-metrics-certs\") pod \"network-metrics-daemon-98x74\" (UID: \"229213f7-7a25-4982-a449-0122912ba05c\") " pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:55:42 crc kubenswrapper[4896]: E0218 14:55:42.114367 4896 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 14:55:42 crc kubenswrapper[4896]: E0218 14:55:42.114458 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/229213f7-7a25-4982-a449-0122912ba05c-metrics-certs podName:229213f7-7a25-4982-a449-0122912ba05c nodeName:}" failed. No retries permitted until 2026-02-18 14:55:46.114433732 +0000 UTC m=+45.043767955 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/229213f7-7a25-4982-a449-0122912ba05c-metrics-certs") pod "network-metrics-daemon-98x74" (UID: "229213f7-7a25-4982-a449-0122912ba05c") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.119013 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:42Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.133883 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:42Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.155065 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54aecfeba56baae4b078096062414aca0ae6e1769fbbc971a15b1e84bfac0538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54aecfeba56baae4b078096062414aca0ae6e1769fbbc971a15b1e84bfac0538\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"message\\\":\\\"umn _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0218 14:55:37.429578 6353 ovnkube.go:599] Stopped ovnkube\\\\nI0218 14:55:37.429657 6353 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0218 14:55:37.429778 6353 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-w862c_openshift-ovn-kubernetes(449d40af-fe23-4350-b66f-de4ce6614177)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:42Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.161495 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.161541 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.161553 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.161570 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.161583 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:42Z","lastTransitionTime":"2026-02-18T14:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.166334 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:42Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.182803 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6e1900020612d617118e0cca524ffaafe1fac28997fcf1f73a75fe743b4f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:42Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.194355 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6wprd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8896bb9-c6fe-4111-a49a-9dadf23e6855\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80d2101411bb762554b416ff237ce9ce2455e43c3c5379118322264ffd9dcd94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5jc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6wprd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:42Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.206852 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:42Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.263477 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.263511 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.263522 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.263540 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.263552 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:42Z","lastTransitionTime":"2026-02-18T14:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.366488 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.366535 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.366544 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.366558 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.366570 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:42Z","lastTransitionTime":"2026-02-18T14:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.468838 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.468881 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.468891 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.468905 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.468915 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:42Z","lastTransitionTime":"2026-02-18T14:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.572386 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.572436 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.572447 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.572464 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.572474 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:42Z","lastTransitionTime":"2026-02-18T14:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.675048 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.675100 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.675113 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.675130 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.675142 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:42Z","lastTransitionTime":"2026-02-18T14:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.778382 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.778432 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.778443 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.778471 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.778491 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:42Z","lastTransitionTime":"2026-02-18T14:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.880819 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.880893 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.880925 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.880942 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.880954 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:42Z","lastTransitionTime":"2026-02-18T14:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.883341 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 21:46:07.515121547 +0000 UTC Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.926973 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.927032 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:55:42 crc kubenswrapper[4896]: E0218 14:55:42.927122 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.927208 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:55:42 crc kubenswrapper[4896]: E0218 14:55:42.927377 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:55:42 crc kubenswrapper[4896]: E0218 14:55:42.927552 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.983975 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.984042 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.984054 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.984076 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:42 crc kubenswrapper[4896]: I0218 14:55:42.984090 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:42Z","lastTransitionTime":"2026-02-18T14:55:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.087356 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.087396 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.087407 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.087422 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.087435 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:43Z","lastTransitionTime":"2026-02-18T14:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.190790 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.190892 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.190905 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.190930 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.190945 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:43Z","lastTransitionTime":"2026-02-18T14:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.293892 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.293922 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.293932 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.293945 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.293953 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:43Z","lastTransitionTime":"2026-02-18T14:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.396350 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.396385 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.396392 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.396405 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.396414 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:43Z","lastTransitionTime":"2026-02-18T14:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.499532 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.499589 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.499628 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.499658 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.499678 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:43Z","lastTransitionTime":"2026-02-18T14:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.602987 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.603051 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.603072 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.603101 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.603125 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:43Z","lastTransitionTime":"2026-02-18T14:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.707049 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.707117 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.707137 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.707168 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.707191 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:43Z","lastTransitionTime":"2026-02-18T14:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.809734 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.809831 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.809861 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.809897 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.809924 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:43Z","lastTransitionTime":"2026-02-18T14:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.884447 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 15:37:30.657198923 +0000 UTC Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.913647 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.913744 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.913768 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.913800 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.913825 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:43Z","lastTransitionTime":"2026-02-18T14:55:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:43 crc kubenswrapper[4896]: I0218 14:55:43.927349 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:55:43 crc kubenswrapper[4896]: E0218 14:55:43.927581 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.017131 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.017219 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.017264 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.017287 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.017299 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:44Z","lastTransitionTime":"2026-02-18T14:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.120381 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.120484 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.120505 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.120538 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.120564 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:44Z","lastTransitionTime":"2026-02-18T14:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.223328 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.223380 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.223396 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.223417 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.223431 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:44Z","lastTransitionTime":"2026-02-18T14:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.326621 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.326680 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.326699 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.326727 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.326746 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:44Z","lastTransitionTime":"2026-02-18T14:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.430381 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.430425 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.430435 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.430455 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.430465 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:44Z","lastTransitionTime":"2026-02-18T14:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.533582 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.533631 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.533647 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.533673 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.533685 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:44Z","lastTransitionTime":"2026-02-18T14:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.636731 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.637075 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.637171 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.637287 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.637384 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:44Z","lastTransitionTime":"2026-02-18T14:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.741201 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.741644 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.741774 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.741867 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.742003 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:44Z","lastTransitionTime":"2026-02-18T14:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.845355 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.845758 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.845874 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.846021 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.846127 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:44Z","lastTransitionTime":"2026-02-18T14:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.885131 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 09:28:06.740027026 +0000 UTC Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.927679 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.927780 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:55:44 crc kubenswrapper[4896]: E0218 14:55:44.927932 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.927698 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:55:44 crc kubenswrapper[4896]: E0218 14:55:44.928114 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:55:44 crc kubenswrapper[4896]: E0218 14:55:44.928295 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.948291 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.948325 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.948334 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.948350 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:44 crc kubenswrapper[4896]: I0218 14:55:44.948359 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:44Z","lastTransitionTime":"2026-02-18T14:55:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.050807 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.050844 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.050854 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.050903 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.050917 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:45Z","lastTransitionTime":"2026-02-18T14:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.153344 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.153384 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.153396 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.153410 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.153419 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:45Z","lastTransitionTime":"2026-02-18T14:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.255865 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.255919 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.255943 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.255969 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.255990 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:45Z","lastTransitionTime":"2026-02-18T14:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.359165 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.359201 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.359214 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.359252 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.359266 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:45Z","lastTransitionTime":"2026-02-18T14:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.461321 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.461380 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.461389 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.461404 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.461414 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:45Z","lastTransitionTime":"2026-02-18T14:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.564725 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.564764 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.564772 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.564786 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.564795 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:45Z","lastTransitionTime":"2026-02-18T14:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.667746 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.667788 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.667800 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.667815 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.667825 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:45Z","lastTransitionTime":"2026-02-18T14:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.771306 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.771344 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.771354 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.771368 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.771379 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:45Z","lastTransitionTime":"2026-02-18T14:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.874400 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.874434 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.874443 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.874457 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.874468 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:45Z","lastTransitionTime":"2026-02-18T14:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.885915 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 02:11:44.159487511 +0000 UTC Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.927357 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:55:45 crc kubenswrapper[4896]: E0218 14:55:45.927459 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.976669 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.976723 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.976740 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.976757 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:45 crc kubenswrapper[4896]: I0218 14:55:45.976770 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:45Z","lastTransitionTime":"2026-02-18T14:55:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.079203 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.079263 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.079275 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.079292 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.079307 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:46Z","lastTransitionTime":"2026-02-18T14:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.155473 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/229213f7-7a25-4982-a449-0122912ba05c-metrics-certs\") pod \"network-metrics-daemon-98x74\" (UID: \"229213f7-7a25-4982-a449-0122912ba05c\") " pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:55:46 crc kubenswrapper[4896]: E0218 14:55:46.155612 4896 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 14:55:46 crc kubenswrapper[4896]: E0218 14:55:46.155977 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/229213f7-7a25-4982-a449-0122912ba05c-metrics-certs podName:229213f7-7a25-4982-a449-0122912ba05c nodeName:}" failed. No retries permitted until 2026-02-18 14:55:54.155957237 +0000 UTC m=+53.085291450 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/229213f7-7a25-4982-a449-0122912ba05c-metrics-certs") pod "network-metrics-daemon-98x74" (UID: "229213f7-7a25-4982-a449-0122912ba05c") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.181494 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.181528 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.181536 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.181550 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.181563 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:46Z","lastTransitionTime":"2026-02-18T14:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.284028 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.284061 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.284069 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.284082 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.284092 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:46Z","lastTransitionTime":"2026-02-18T14:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.386266 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.386515 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.386593 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.386711 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.386788 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:46Z","lastTransitionTime":"2026-02-18T14:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.489440 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.489574 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.489595 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.489622 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.489638 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:46Z","lastTransitionTime":"2026-02-18T14:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.592985 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.593041 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.593052 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.593072 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.593086 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:46Z","lastTransitionTime":"2026-02-18T14:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.696528 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.696592 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.696605 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.696635 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.696651 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:46Z","lastTransitionTime":"2026-02-18T14:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.798887 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.798942 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.798951 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.798962 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.798970 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:46Z","lastTransitionTime":"2026-02-18T14:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.886483 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 15:19:39.994403335 +0000 UTC Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.901483 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.901542 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.901557 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.901578 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.901592 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:46Z","lastTransitionTime":"2026-02-18T14:55:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.927964 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.928010 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:55:46 crc kubenswrapper[4896]: I0218 14:55:46.928023 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:55:46 crc kubenswrapper[4896]: E0218 14:55:46.928149 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:55:46 crc kubenswrapper[4896]: E0218 14:55:46.928313 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:55:46 crc kubenswrapper[4896]: E0218 14:55:46.928596 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.004168 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.004210 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.004222 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.004267 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.004281 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:47Z","lastTransitionTime":"2026-02-18T14:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.107329 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.107374 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.107385 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.107403 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.107414 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:47Z","lastTransitionTime":"2026-02-18T14:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.210065 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.210149 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.210174 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.210220 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.210335 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:47Z","lastTransitionTime":"2026-02-18T14:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.314206 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.314314 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.314329 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.314348 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.314361 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:47Z","lastTransitionTime":"2026-02-18T14:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.417225 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.417321 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.417341 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.417369 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.417389 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:47Z","lastTransitionTime":"2026-02-18T14:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.519784 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.520099 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.520294 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.520454 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.520626 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:47Z","lastTransitionTime":"2026-02-18T14:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.623571 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.623657 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.623669 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.623687 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.623700 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:47Z","lastTransitionTime":"2026-02-18T14:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.726213 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.726539 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.726670 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.726772 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.726854 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:47Z","lastTransitionTime":"2026-02-18T14:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.829604 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.829824 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.830014 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.830097 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.830188 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:47Z","lastTransitionTime":"2026-02-18T14:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.887512 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 20:45:07.64311764 +0000 UTC Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.927198 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:55:47 crc kubenswrapper[4896]: E0218 14:55:47.927646 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.931858 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.931987 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.932094 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.932174 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:47 crc kubenswrapper[4896]: I0218 14:55:47.932286 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:47Z","lastTransitionTime":"2026-02-18T14:55:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.034812 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.035017 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.035103 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.035193 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.035274 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:48Z","lastTransitionTime":"2026-02-18T14:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.137677 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.137916 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.137986 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.138060 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.138149 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:48Z","lastTransitionTime":"2026-02-18T14:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.239788 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.239829 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.239839 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.239854 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.239865 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:48Z","lastTransitionTime":"2026-02-18T14:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.342338 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.342801 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.343044 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.343358 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.343583 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:48Z","lastTransitionTime":"2026-02-18T14:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.446648 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.446702 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.446721 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.446740 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.446753 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:48Z","lastTransitionTime":"2026-02-18T14:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.549417 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.549450 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.549459 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.549472 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.549482 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:48Z","lastTransitionTime":"2026-02-18T14:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.651847 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.652147 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.652228 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.652332 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.652393 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:48Z","lastTransitionTime":"2026-02-18T14:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.754910 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.755121 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.755186 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.755306 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.755373 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:48Z","lastTransitionTime":"2026-02-18T14:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.857801 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.857866 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.857880 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.857895 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.857908 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:48Z","lastTransitionTime":"2026-02-18T14:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.888651 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 07:09:49.461586577 +0000 UTC Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.927961 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.928003 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:55:48 crc kubenswrapper[4896]: E0218 14:55:48.928098 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.927961 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:55:48 crc kubenswrapper[4896]: E0218 14:55:48.928193 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:55:48 crc kubenswrapper[4896]: E0218 14:55:48.928290 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.960095 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.960127 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.960135 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.960149 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:48 crc kubenswrapper[4896]: I0218 14:55:48.960158 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:48Z","lastTransitionTime":"2026-02-18T14:55:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.062438 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.062472 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.062480 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.062492 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.062501 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:49Z","lastTransitionTime":"2026-02-18T14:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.164673 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.164710 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.164719 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.164733 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.164743 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:49Z","lastTransitionTime":"2026-02-18T14:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.266163 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.266226 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.266259 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.266275 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.266284 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:49Z","lastTransitionTime":"2026-02-18T14:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.339948 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.339980 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.339991 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.340005 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.340015 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:49Z","lastTransitionTime":"2026-02-18T14:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:49 crc kubenswrapper[4896]: E0218 14:55:49.354863 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3d8f65ef-15b5-4cc3-be66-c8dd6cb956e9\\\",\\\"systemUUID\\\":\\\"d255e81f-4350-4697-bbdf-3f95dec57eda\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:49Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.358421 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.358550 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.358785 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.358963 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.359134 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:49Z","lastTransitionTime":"2026-02-18T14:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:49 crc kubenswrapper[4896]: E0218 14:55:49.370485 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3d8f65ef-15b5-4cc3-be66-c8dd6cb956e9\\\",\\\"systemUUID\\\":\\\"d255e81f-4350-4697-bbdf-3f95dec57eda\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:49Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.377037 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.377089 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.377110 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.377137 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.377157 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:49Z","lastTransitionTime":"2026-02-18T14:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:49 crc kubenswrapper[4896]: E0218 14:55:49.391846 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3d8f65ef-15b5-4cc3-be66-c8dd6cb956e9\\\",\\\"systemUUID\\\":\\\"d255e81f-4350-4697-bbdf-3f95dec57eda\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:49Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.397058 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.397118 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.397135 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.397160 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.397179 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:49Z","lastTransitionTime":"2026-02-18T14:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:49 crc kubenswrapper[4896]: E0218 14:55:49.409468 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3d8f65ef-15b5-4cc3-be66-c8dd6cb956e9\\\",\\\"systemUUID\\\":\\\"d255e81f-4350-4697-bbdf-3f95dec57eda\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:49Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.416967 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.416992 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.417000 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.417113 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.417125 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:49Z","lastTransitionTime":"2026-02-18T14:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:49 crc kubenswrapper[4896]: E0218 14:55:49.433742 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3d8f65ef-15b5-4cc3-be66-c8dd6cb956e9\\\",\\\"systemUUID\\\":\\\"d255e81f-4350-4697-bbdf-3f95dec57eda\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:49Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:49 crc kubenswrapper[4896]: E0218 14:55:49.433976 4896 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.435511 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.435563 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.435579 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.435600 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.435618 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:49Z","lastTransitionTime":"2026-02-18T14:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.538340 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.538410 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.538429 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.538453 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.538476 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:49Z","lastTransitionTime":"2026-02-18T14:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.641258 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.641299 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.641314 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.641334 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.641350 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:49Z","lastTransitionTime":"2026-02-18T14:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.744105 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.744166 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.744181 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.744204 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.744217 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:49Z","lastTransitionTime":"2026-02-18T14:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.846955 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.846997 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.847009 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.847024 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.847034 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:49Z","lastTransitionTime":"2026-02-18T14:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.889553 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 18:16:27.806529891 +0000 UTC Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.927514 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:55:49 crc kubenswrapper[4896]: E0218 14:55:49.927671 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.948907 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.948951 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.948998 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.949015 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:49 crc kubenswrapper[4896]: I0218 14:55:49.949026 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:49Z","lastTransitionTime":"2026-02-18T14:55:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.051825 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.051853 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.051861 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.051876 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.051884 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:50Z","lastTransitionTime":"2026-02-18T14:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.154163 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.154209 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.154221 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.154265 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.154276 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:50Z","lastTransitionTime":"2026-02-18T14:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.256630 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.256883 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.257148 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.257577 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.260212 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:50Z","lastTransitionTime":"2026-02-18T14:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.363069 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.363130 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.363146 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.363173 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.363189 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:50Z","lastTransitionTime":"2026-02-18T14:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.465583 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.465611 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.465620 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.465632 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.465639 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:50Z","lastTransitionTime":"2026-02-18T14:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.567758 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.567816 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.567838 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.567865 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.567882 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:50Z","lastTransitionTime":"2026-02-18T14:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.671472 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.671772 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.671955 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.672149 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.672373 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:50Z","lastTransitionTime":"2026-02-18T14:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.775588 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.775932 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.776056 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.776178 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.776357 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:50Z","lastTransitionTime":"2026-02-18T14:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.879319 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.879382 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.879398 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.879423 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.879439 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:50Z","lastTransitionTime":"2026-02-18T14:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.889972 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 20:23:41.416219828 +0000 UTC Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.927577 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.927668 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:55:50 crc kubenswrapper[4896]: E0218 14:55:50.928008 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:55:50 crc kubenswrapper[4896]: E0218 14:55:50.928007 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.927703 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:55:50 crc kubenswrapper[4896]: E0218 14:55:50.928084 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.982203 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.982286 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.982306 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.982329 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:50 crc kubenswrapper[4896]: I0218 14:55:50.982347 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:50Z","lastTransitionTime":"2026-02-18T14:55:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.085036 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.085090 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.085098 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.085111 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.085121 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:51Z","lastTransitionTime":"2026-02-18T14:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.188261 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.188294 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.188304 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.188318 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.188329 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:51Z","lastTransitionTime":"2026-02-18T14:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.290079 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.290113 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.290123 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.290137 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.290146 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:51Z","lastTransitionTime":"2026-02-18T14:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.392882 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.392927 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.392939 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.392956 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.392965 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:51Z","lastTransitionTime":"2026-02-18T14:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.495364 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.495389 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.495397 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.495409 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.495417 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:51Z","lastTransitionTime":"2026-02-18T14:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.598069 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.598123 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.598140 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.598166 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.598183 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:51Z","lastTransitionTime":"2026-02-18T14:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.700514 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.700555 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.700564 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.700581 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.700594 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:51Z","lastTransitionTime":"2026-02-18T14:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.803315 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.803397 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.803422 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.803452 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.803472 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:51Z","lastTransitionTime":"2026-02-18T14:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.890785 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 17:52:13.109730145 +0000 UTC Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.906186 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.906301 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.906331 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.906361 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.906385 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:51Z","lastTransitionTime":"2026-02-18T14:55:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.928011 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:55:51 crc kubenswrapper[4896]: E0218 14:55:51.928316 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.949316 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:51Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.971142 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:51Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:51 crc kubenswrapper[4896]: I0218 14:55:51.988669 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32bd695f-cf0b-425a-b764-87b3f678d4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://390187c9005f214d18d43cd6b4f987ccc7d2197f8629997ca4c5cd3ffb49c777\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7021f783fefb51e5fb6b701e58eb9279a0513b9ff37cba435d69e1b7bb1d1a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8j9gb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:51Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.001059 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-98x74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"229213f7-7a25-4982-a449-0122912ba05c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-98x74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:51Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.008282 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.008316 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.008324 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.008337 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.008346 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:52Z","lastTransitionTime":"2026-02-18T14:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.016670 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:52Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.032362 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:52Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.047004 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:52Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.067286 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:52Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.089454 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://54aecfeba56baae4b078096062414aca0ae6e1769fbbc971a15b1e84bfac0538\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54aecfeba56baae4b078096062414aca0ae6e1769fbbc971a15b1e84bfac0538\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"message\\\":\\\"umn _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0218 14:55:37.429578 6353 ovnkube.go:599] Stopped ovnkube\\\\nI0218 14:55:37.429657 6353 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0218 14:55:37.429778 6353 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-w862c_openshift-ovn-kubernetes(449d40af-fe23-4350-b66f-de4ce6614177)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:52Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.103066 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6e1900020612d617118e0cca524ffaafe1fac28997fcf1f73a75fe743b4f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:52Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.111722 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.111757 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.111765 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.111779 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.111787 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:52Z","lastTransitionTime":"2026-02-18T14:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.115831 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6wprd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8896bb9-c6fe-4111-a49a-9dadf23e6855\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80d2101411bb762554b416ff237ce9ce2455e43c3c5379118322264ffd9dcd94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5jc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6wprd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:52Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.126910 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:52Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.138358 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:52Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.152130 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:52Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.162897 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:52Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.181624 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:52Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.193689 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e1cf96a28a07565ec2d56a4ffc32c10e95356b7ebf2399b6b3ba833274a68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:52Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.214313 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.214355 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.214370 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.214407 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.214424 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:52Z","lastTransitionTime":"2026-02-18T14:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.316548 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.316576 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.316587 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.316602 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.316620 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:52Z","lastTransitionTime":"2026-02-18T14:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.418745 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.418790 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.418798 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.418811 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.418819 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:52Z","lastTransitionTime":"2026-02-18T14:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.521183 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.521226 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.521263 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.521287 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.521299 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:52Z","lastTransitionTime":"2026-02-18T14:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.623090 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.623510 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.623681 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.623813 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.623932 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:52Z","lastTransitionTime":"2026-02-18T14:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.726294 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.726538 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.726608 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.726699 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.726762 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:52Z","lastTransitionTime":"2026-02-18T14:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.829378 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.829410 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.829423 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.829441 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.829453 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:52Z","lastTransitionTime":"2026-02-18T14:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.891359 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 05:15:05.892311075 +0000 UTC Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.926972 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.926972 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.926995 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:55:52 crc kubenswrapper[4896]: E0218 14:55:52.927085 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:55:52 crc kubenswrapper[4896]: E0218 14:55:52.927732 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:55:52 crc kubenswrapper[4896]: E0218 14:55:52.927833 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.928222 4896 scope.go:117] "RemoveContainer" containerID="54aecfeba56baae4b078096062414aca0ae6e1769fbbc971a15b1e84bfac0538" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.931001 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.931042 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.931056 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.931075 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:52 crc kubenswrapper[4896]: I0218 14:55:52.931090 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:52Z","lastTransitionTime":"2026-02-18T14:55:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.033313 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.033363 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.033374 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.033392 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.033405 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:53Z","lastTransitionTime":"2026-02-18T14:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.135695 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.135748 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.135766 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.135783 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.135796 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:53Z","lastTransitionTime":"2026-02-18T14:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.238399 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.238431 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.238439 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.238453 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.238463 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:53Z","lastTransitionTime":"2026-02-18T14:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.255139 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w862c_449d40af-fe23-4350-b66f-de4ce6614177/ovnkube-controller/1.log" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.258410 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" event={"ID":"449d40af-fe23-4350-b66f-de4ce6614177","Type":"ContainerStarted","Data":"3705371cdc4101eed891047c7e68b95155ea083842e0343ab98763e25ea00f73"} Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.258976 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.276821 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:53Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.299274 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3705371cdc4101eed891047c7e68b95155ea083842e0343ab98763e25ea00f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54aecfeba56baae4b078096062414aca0ae6e1769fbbc971a15b1e84bfac0538\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"message\\\":\\\"umn _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0218 14:55:37.429578 6353 ovnkube.go:599] Stopped ovnkube\\\\nI0218 14:55:37.429657 6353 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0218 14:55:37.429778 6353 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:53Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.315660 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:53Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.332941 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:53Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.341010 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.341150 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.341258 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.341351 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.341441 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:53Z","lastTransitionTime":"2026-02-18T14:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.353829 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6e1900020612d617118e0cca524ffaafe1fac28997fcf1f73a75fe743b4f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:53Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.370535 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6wprd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8896bb9-c6fe-4111-a49a-9dadf23e6855\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80d2101411bb762554b416ff237ce9ce2455e43c3c5379118322264ffd9dcd94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5jc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6wprd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:53Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.394184 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:53Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.421530 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e1cf96a28a07565ec2d56a4ffc32c10e95356b7ebf2399b6b3ba833274a68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:53Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.445159 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.445213 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.445232 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.445281 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.445298 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:53Z","lastTransitionTime":"2026-02-18T14:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.451381 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:53Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.465077 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:53Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.505715 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-98x74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"229213f7-7a25-4982-a449-0122912ba05c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-98x74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:53Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.527174 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:53Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.540507 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:53Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.547900 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.547943 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.547956 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.547974 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.547986 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:53Z","lastTransitionTime":"2026-02-18T14:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.554917 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:53Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.567953 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:53Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.579225 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:53Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.591673 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32bd695f-cf0b-425a-b764-87b3f678d4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://390187c9005f214d18d43cd6b4f987ccc7d2197f8629997ca4c5cd3ffb49c777\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7021f783fefb51e5fb6b701e58eb9279a0513b9ff37cba435d69e1b7bb1d1a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8j9gb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:53Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.650179 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.650217 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.650227 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.650254 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.650264 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:53Z","lastTransitionTime":"2026-02-18T14:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.751805 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.751838 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.751846 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.751859 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.751868 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:53Z","lastTransitionTime":"2026-02-18T14:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.854121 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.854176 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.854188 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.854208 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.854221 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:53Z","lastTransitionTime":"2026-02-18T14:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.892521 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 18:29:36.111234347 +0000 UTC Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.927214 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:55:53 crc kubenswrapper[4896]: E0218 14:55:53.927649 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.956567 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.956596 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.956605 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.956634 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:53 crc kubenswrapper[4896]: I0218 14:55:53.956644 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:53Z","lastTransitionTime":"2026-02-18T14:55:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.059300 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.059351 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.059369 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.059390 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.059405 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:54Z","lastTransitionTime":"2026-02-18T14:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.162829 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.162859 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.162868 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.162881 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.162892 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:54Z","lastTransitionTime":"2026-02-18T14:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.236761 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/229213f7-7a25-4982-a449-0122912ba05c-metrics-certs\") pod \"network-metrics-daemon-98x74\" (UID: \"229213f7-7a25-4982-a449-0122912ba05c\") " pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:55:54 crc kubenswrapper[4896]: E0218 14:55:54.236896 4896 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 14:55:54 crc kubenswrapper[4896]: E0218 14:55:54.236944 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/229213f7-7a25-4982-a449-0122912ba05c-metrics-certs podName:229213f7-7a25-4982-a449-0122912ba05c nodeName:}" failed. No retries permitted until 2026-02-18 14:56:10.236930345 +0000 UTC m=+69.166264558 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/229213f7-7a25-4982-a449-0122912ba05c-metrics-certs") pod "network-metrics-daemon-98x74" (UID: "229213f7-7a25-4982-a449-0122912ba05c") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.264802 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.264847 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.264863 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.264885 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.264903 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:54Z","lastTransitionTime":"2026-02-18T14:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.265199 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w862c_449d40af-fe23-4350-b66f-de4ce6614177/ovnkube-controller/2.log" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.266067 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w862c_449d40af-fe23-4350-b66f-de4ce6614177/ovnkube-controller/1.log" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.269037 4896 generic.go:334] "Generic (PLEG): container finished" podID="449d40af-fe23-4350-b66f-de4ce6614177" containerID="3705371cdc4101eed891047c7e68b95155ea083842e0343ab98763e25ea00f73" exitCode=1 Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.269100 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" event={"ID":"449d40af-fe23-4350-b66f-de4ce6614177","Type":"ContainerDied","Data":"3705371cdc4101eed891047c7e68b95155ea083842e0343ab98763e25ea00f73"} Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.269143 4896 scope.go:117] "RemoveContainer" containerID="54aecfeba56baae4b078096062414aca0ae6e1769fbbc971a15b1e84bfac0538" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.270025 4896 scope.go:117] "RemoveContainer" containerID="3705371cdc4101eed891047c7e68b95155ea083842e0343ab98763e25ea00f73" Feb 18 14:55:54 crc kubenswrapper[4896]: E0218 14:55:54.270155 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-w862c_openshift-ovn-kubernetes(449d40af-fe23-4350-b66f-de4ce6614177)\"" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" podUID="449d40af-fe23-4350-b66f-de4ce6614177" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.286815 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:54Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.304527 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32bd695f-cf0b-425a-b764-87b3f678d4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://390187c9005f214d18d43cd6b4f987ccc7d2197f8629997ca4c5cd3ffb49c777\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7021f783fefb51e5fb6b701e58eb9279a0513b9ff37cba435d69e1b7bb1d1a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8j9gb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:54Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.318967 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-98x74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"229213f7-7a25-4982-a449-0122912ba05c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-98x74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:54Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.335831 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:54Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.349700 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:54Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.364048 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:54Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.367188 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.367272 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.367332 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.367397 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.367417 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:54Z","lastTransitionTime":"2026-02-18T14:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.378381 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:54Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.391481 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:54Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.410818 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3705371cdc4101eed891047c7e68b95155ea083842e0343ab98763e25ea00f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://54aecfeba56baae4b078096062414aca0ae6e1769fbbc971a15b1e84bfac0538\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"message\\\":\\\"umn _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:dce28c51-c9f1-478b-97c8-7e209d6e7cbe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0218 14:55:37.429578 6353 ovnkube.go:599] Stopped ovnkube\\\\nI0218 14:55:37.429657 6353 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0218 14:55:37.429778 6353 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3705371cdc4101eed891047c7e68b95155ea083842e0343ab98763e25ea00f73\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:55:53Z\\\",\\\"message\\\":\\\"ew object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0218 14:55:53.908904 6577 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nI0218 14:55:53.908922 6577 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0218 14:55:53.908938 6577 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI0218 14:55:53.908947 6577 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}\\\\nI0218 14:55:53.908958 6577 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 2.740953ms\\\\nI0218 14:55:53.908969 6577 services_controller.go:356] Processing sync for service openshift-network-diagnostics/network-check-target for network=default\\\\nF0218 14:55:53.908970 6577 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:54Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.421719 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6wprd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8896bb9-c6fe-4111-a49a-9dadf23e6855\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80d2101411bb762554b416ff237ce9ce2455e43c3c5379118322264ffd9dcd94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5jc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6wprd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:54Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.434739 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:54Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.447160 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:54Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.461442 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6e1900020612d617118e0cca524ffaafe1fac28997fcf1f73a75fe743b4f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:54Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.469959 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.470016 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.470029 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.470047 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.470089 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:54Z","lastTransitionTime":"2026-02-18T14:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.472327 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:54Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.496933 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:54Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.508352 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e1cf96a28a07565ec2d56a4ffc32c10e95356b7ebf2399b6b3ba833274a68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:54Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.519910 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:54Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.571635 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.571675 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.571700 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.571713 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.571721 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:54Z","lastTransitionTime":"2026-02-18T14:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.673786 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.673820 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.673831 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.673847 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.673858 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:54Z","lastTransitionTime":"2026-02-18T14:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.742723 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.742875 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:55:54 crc kubenswrapper[4896]: E0218 14:55:54.742944 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:56:26.742914011 +0000 UTC m=+85.672248234 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.743006 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:55:54 crc kubenswrapper[4896]: E0218 14:55:54.743029 4896 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.743051 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:55:54 crc kubenswrapper[4896]: E0218 14:55:54.743102 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 14:56:26.743079046 +0000 UTC m=+85.672413289 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.743137 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:55:54 crc kubenswrapper[4896]: E0218 14:55:54.743142 4896 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 14:55:54 crc kubenswrapper[4896]: E0218 14:55:54.743212 4896 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 14:55:54 crc kubenswrapper[4896]: E0218 14:55:54.743227 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 14:56:26.743214579 +0000 UTC m=+85.672548832 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 14:55:54 crc kubenswrapper[4896]: E0218 14:55:54.743259 4896 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 14:55:54 crc kubenswrapper[4896]: E0218 14:55:54.743312 4896 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 14:55:54 crc kubenswrapper[4896]: E0218 14:55:54.743268 4896 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 14:55:54 crc kubenswrapper[4896]: E0218 14:55:54.743369 4896 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 14:55:54 crc kubenswrapper[4896]: E0218 14:55:54.743385 4896 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 14:55:54 crc kubenswrapper[4896]: E0218 14:55:54.743369 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-18 14:56:26.743353663 +0000 UTC m=+85.672687906 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 14:55:54 crc kubenswrapper[4896]: E0218 14:55:54.743443 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-18 14:56:26.743432115 +0000 UTC m=+85.672766418 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.776311 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.776359 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.776404 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.776422 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.776432 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:54Z","lastTransitionTime":"2026-02-18T14:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.878559 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.878597 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.878608 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.878622 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.878631 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:54Z","lastTransitionTime":"2026-02-18T14:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.893047 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 07:50:45.727910602 +0000 UTC Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.927111 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.927144 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.927172 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:55:54 crc kubenswrapper[4896]: E0218 14:55:54.927280 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:55:54 crc kubenswrapper[4896]: E0218 14:55:54.927335 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:55:54 crc kubenswrapper[4896]: E0218 14:55:54.927399 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.980636 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.980669 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.980685 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.980702 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:54 crc kubenswrapper[4896]: I0218 14:55:54.980714 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:54Z","lastTransitionTime":"2026-02-18T14:55:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.082347 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.082387 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.082395 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.082408 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.082418 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:55Z","lastTransitionTime":"2026-02-18T14:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.185406 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.185465 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.185482 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.185504 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.185520 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:55Z","lastTransitionTime":"2026-02-18T14:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.274875 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w862c_449d40af-fe23-4350-b66f-de4ce6614177/ovnkube-controller/2.log" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.280809 4896 scope.go:117] "RemoveContainer" containerID="3705371cdc4101eed891047c7e68b95155ea083842e0343ab98763e25ea00f73" Feb 18 14:55:55 crc kubenswrapper[4896]: E0218 14:55:55.281040 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-w862c_openshift-ovn-kubernetes(449d40af-fe23-4350-b66f-de4ce6614177)\"" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" podUID="449d40af-fe23-4350-b66f-de4ce6614177" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.287994 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.288066 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.288075 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.288089 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.288100 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:55Z","lastTransitionTime":"2026-02-18T14:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.300214 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:55Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.317590 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:55Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.335094 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:55Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.352158 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:55Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.363956 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:55Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.379524 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32bd695f-cf0b-425a-b764-87b3f678d4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://390187c9005f214d18d43cd6b4f987ccc7d2197f8629997ca4c5cd3ffb49c777\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7021f783fefb51e5fb6b701e58eb9279a0513b9ff37cba435d69e1b7bb1d1a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8j9gb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:55Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.390109 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.390138 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.390149 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.390164 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.390175 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:55Z","lastTransitionTime":"2026-02-18T14:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.393983 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-98x74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"229213f7-7a25-4982-a449-0122912ba05c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-98x74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:55Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.407343 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:55Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.429801 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3705371cdc4101eed891047c7e68b95155ea083842e0343ab98763e25ea00f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3705371cdc4101eed891047c7e68b95155ea083842e0343ab98763e25ea00f73\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:55:53Z\\\",\\\"message\\\":\\\"ew object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0218 14:55:53.908904 6577 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nI0218 14:55:53.908922 6577 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0218 14:55:53.908938 6577 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI0218 14:55:53.908947 6577 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}\\\\nI0218 14:55:53.908958 6577 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 2.740953ms\\\\nI0218 14:55:53.908969 6577 services_controller.go:356] Processing sync for service openshift-network-diagnostics/network-check-target for network=default\\\\nF0218 14:55:53.908970 6577 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-w862c_openshift-ovn-kubernetes(449d40af-fe23-4350-b66f-de4ce6614177)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:55Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.441501 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:55Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.453589 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:55Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.468392 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6e1900020612d617118e0cca524ffaafe1fac28997fcf1f73a75fe743b4f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:55Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.480689 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6wprd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8896bb9-c6fe-4111-a49a-9dadf23e6855\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80d2101411bb762554b416ff237ce9ce2455e43c3c5379118322264ffd9dcd94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5jc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6wprd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:55Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.498145 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.498260 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.498281 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.498317 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.498391 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:55Z","lastTransitionTime":"2026-02-18T14:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.506386 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:55Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.521387 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e1cf96a28a07565ec2d56a4ffc32c10e95356b7ebf2399b6b3ba833274a68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:55Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.539525 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:55Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.555772 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:55Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.601354 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.601425 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.601438 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.601483 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.601502 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:55Z","lastTransitionTime":"2026-02-18T14:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.704014 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.704041 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.704049 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.704061 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.704068 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:55Z","lastTransitionTime":"2026-02-18T14:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.806079 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.806121 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.806135 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.806154 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.806169 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:55Z","lastTransitionTime":"2026-02-18T14:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.893788 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 07:02:00.048287591 +0000 UTC Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.908606 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.908653 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.908667 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.908689 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.908704 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:55Z","lastTransitionTime":"2026-02-18T14:55:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:55 crc kubenswrapper[4896]: I0218 14:55:55.927362 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:55:55 crc kubenswrapper[4896]: E0218 14:55:55.927488 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.011077 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.011114 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.011123 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.011138 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.011149 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:56Z","lastTransitionTime":"2026-02-18T14:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.113975 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.114007 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.114015 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.114030 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.114038 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:56Z","lastTransitionTime":"2026-02-18T14:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.216335 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.216501 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.216518 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.216533 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.216545 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:56Z","lastTransitionTime":"2026-02-18T14:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.318839 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.318875 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.318885 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.318899 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.318910 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:56Z","lastTransitionTime":"2026-02-18T14:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.421265 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.421310 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.421318 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.421333 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.421341 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:56Z","lastTransitionTime":"2026-02-18T14:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.524047 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.524131 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.524149 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.524170 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.524184 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:56Z","lastTransitionTime":"2026-02-18T14:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.626449 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.626477 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.626486 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.626500 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.626509 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:56Z","lastTransitionTime":"2026-02-18T14:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.729108 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.729142 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.729152 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.729167 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.729178 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:56Z","lastTransitionTime":"2026-02-18T14:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.831919 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.831965 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.831982 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.832002 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.832015 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:56Z","lastTransitionTime":"2026-02-18T14:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.894269 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 06:59:51.545446376 +0000 UTC Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.927912 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.927980 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:55:56 crc kubenswrapper[4896]: E0218 14:55:56.928071 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.927919 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:55:56 crc kubenswrapper[4896]: E0218 14:55:56.928181 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:55:56 crc kubenswrapper[4896]: E0218 14:55:56.928268 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.933919 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.933945 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.933953 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.933966 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:56 crc kubenswrapper[4896]: I0218 14:55:56.933976 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:56Z","lastTransitionTime":"2026-02-18T14:55:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.037062 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.037101 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.037112 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.037131 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.037146 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:57Z","lastTransitionTime":"2026-02-18T14:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.139940 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.139977 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.139986 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.140006 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.140016 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:57Z","lastTransitionTime":"2026-02-18T14:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.242743 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.242828 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.242845 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.242870 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.242895 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:57Z","lastTransitionTime":"2026-02-18T14:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.311746 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.324426 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.345427 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.345459 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.345470 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.345485 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.345497 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:57Z","lastTransitionTime":"2026-02-18T14:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.347668 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.364045 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e1cf96a28a07565ec2d56a4ffc32c10e95356b7ebf2399b6b3ba833274a68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.379801 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.392194 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.403919 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-98x74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"229213f7-7a25-4982-a449-0122912ba05c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-98x74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.422180 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.432551 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.443389 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.447986 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.448030 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.448046 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.448068 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.448081 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:57Z","lastTransitionTime":"2026-02-18T14:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.457275 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.465649 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.481551 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32bd695f-cf0b-425a-b764-87b3f678d4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://390187c9005f214d18d43cd6b4f987ccc7d2197f8629997ca4c5cd3ffb49c777\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7021f783fefb51e5fb6b701e58eb9279a0513b9ff37cba435d69e1b7bb1d1a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8j9gb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.493697 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.513829 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3705371cdc4101eed891047c7e68b95155ea083842e0343ab98763e25ea00f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3705371cdc4101eed891047c7e68b95155ea083842e0343ab98763e25ea00f73\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:55:53Z\\\",\\\"message\\\":\\\"ew object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0218 14:55:53.908904 6577 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nI0218 14:55:53.908922 6577 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0218 14:55:53.908938 6577 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI0218 14:55:53.908947 6577 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}\\\\nI0218 14:55:53.908958 6577 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 2.740953ms\\\\nI0218 14:55:53.908969 6577 services_controller.go:356] Processing sync for service openshift-network-diagnostics/network-check-target for network=default\\\\nF0218 14:55:53.908970 6577 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-w862c_openshift-ovn-kubernetes(449d40af-fe23-4350-b66f-de4ce6614177)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.530400 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.541414 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.550751 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.550785 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.550793 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.550806 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.550816 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:57Z","lastTransitionTime":"2026-02-18T14:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.561480 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6e1900020612d617118e0cca524ffaafe1fac28997fcf1f73a75fe743b4f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.572570 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6wprd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8896bb9-c6fe-4111-a49a-9dadf23e6855\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80d2101411bb762554b416ff237ce9ce2455e43c3c5379118322264ffd9dcd94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5jc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6wprd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.652852 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.652917 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.652926 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.652940 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.652948 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:57Z","lastTransitionTime":"2026-02-18T14:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.690716 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.711757 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.726148 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e1cf96a28a07565ec2d56a4ffc32c10e95356b7ebf2399b6b3ba833274a68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.739788 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"219e676d-1cbc-4503-bef3-7b3c5af4e0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd5e328de9a4e016452b309a56258165a94fdd43c0e26f7444a4a765f3cd11c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec861caaac932e4489277e16914fa252647b155492fec4edeb789ea62e11eac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d26b99304cc56b1e016ba622e79f535b6dfd04c197b698dae1720a1ee44afab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc64cbe4646c598d98e96a7ff130d52867c29c552541b1c7adea6530a582e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bdc64cbe4646c598d98e96a7ff130d52867c29c552541b1c7adea6530a582e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.753690 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.754910 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.754991 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.755009 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.755031 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.755050 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:57Z","lastTransitionTime":"2026-02-18T14:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.764489 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.777458 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-98x74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"229213f7-7a25-4982-a449-0122912ba05c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-98x74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.792400 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.804508 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.816900 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.829540 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.840663 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.851181 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32bd695f-cf0b-425a-b764-87b3f678d4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://390187c9005f214d18d43cd6b4f987ccc7d2197f8629997ca4c5cd3ffb49c777\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7021f783fefb51e5fb6b701e58eb9279a0513b9ff37cba435d69e1b7bb1d1a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8j9gb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.859486 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.859537 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.859552 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.859572 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.859586 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:57Z","lastTransitionTime":"2026-02-18T14:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.862218 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.877756 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3705371cdc4101eed891047c7e68b95155ea083842e0343ab98763e25ea00f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3705371cdc4101eed891047c7e68b95155ea083842e0343ab98763e25ea00f73\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:55:53Z\\\",\\\"message\\\":\\\"ew object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0218 14:55:53.908904 6577 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nI0218 14:55:53.908922 6577 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0218 14:55:53.908938 6577 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI0218 14:55:53.908947 6577 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}\\\\nI0218 14:55:53.908958 6577 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 2.740953ms\\\\nI0218 14:55:53.908969 6577 services_controller.go:356] Processing sync for service openshift-network-diagnostics/network-check-target for network=default\\\\nF0218 14:55:53.908970 6577 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-w862c_openshift-ovn-kubernetes(449d40af-fe23-4350-b66f-de4ce6614177)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.887184 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.894853 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 18:16:40.653857623 +0000 UTC Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.898112 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.910191 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6e1900020612d617118e0cca524ffaafe1fac28997fcf1f73a75fe743b4f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.918768 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6wprd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8896bb9-c6fe-4111-a49a-9dadf23e6855\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80d2101411bb762554b416ff237ce9ce2455e43c3c5379118322264ffd9dcd94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5jc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6wprd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:57Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.926967 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:55:57 crc kubenswrapper[4896]: E0218 14:55:57.927080 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.962272 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.962308 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.962322 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.962340 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:57 crc kubenswrapper[4896]: I0218 14:55:57.962355 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:57Z","lastTransitionTime":"2026-02-18T14:55:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.064283 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.064341 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.064358 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.064381 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.064398 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:58Z","lastTransitionTime":"2026-02-18T14:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.166808 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.166845 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.166853 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.166868 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.166877 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:58Z","lastTransitionTime":"2026-02-18T14:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.269592 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.269643 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.269660 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.269682 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.269697 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:58Z","lastTransitionTime":"2026-02-18T14:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.372769 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.372814 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.372824 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.372841 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.372851 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:58Z","lastTransitionTime":"2026-02-18T14:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.475444 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.475482 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.475490 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.475504 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.475513 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:58Z","lastTransitionTime":"2026-02-18T14:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.577896 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.577948 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.577959 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.577973 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.577982 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:58Z","lastTransitionTime":"2026-02-18T14:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.680368 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.680396 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.680403 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.680415 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.680424 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:58Z","lastTransitionTime":"2026-02-18T14:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.782609 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.782637 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.782645 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.782656 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.782665 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:58Z","lastTransitionTime":"2026-02-18T14:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.884647 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.884680 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.884691 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.884707 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.884718 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:58Z","lastTransitionTime":"2026-02-18T14:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.895858 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 06:21:08.862891015 +0000 UTC Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.927282 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.927294 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.927360 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:55:58 crc kubenswrapper[4896]: E0218 14:55:58.927475 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:55:58 crc kubenswrapper[4896]: E0218 14:55:58.927547 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:55:58 crc kubenswrapper[4896]: E0218 14:55:58.927633 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.986971 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.987007 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.987015 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.987027 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:58 crc kubenswrapper[4896]: I0218 14:55:58.987037 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:58Z","lastTransitionTime":"2026-02-18T14:55:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.088920 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.088964 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.088978 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.088998 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.089013 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:59Z","lastTransitionTime":"2026-02-18T14:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.191704 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.191739 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.191748 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.191797 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.191807 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:59Z","lastTransitionTime":"2026-02-18T14:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.294197 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.294257 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.294268 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.294286 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.294298 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:59Z","lastTransitionTime":"2026-02-18T14:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.397062 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.397096 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.397103 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.397115 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.397125 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:59Z","lastTransitionTime":"2026-02-18T14:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.499779 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.499828 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.499840 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.499856 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.499866 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:59Z","lastTransitionTime":"2026-02-18T14:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.602068 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.602109 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.602120 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.602147 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.602159 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:59Z","lastTransitionTime":"2026-02-18T14:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.654875 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.654925 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.654936 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.654954 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.654966 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:59Z","lastTransitionTime":"2026-02-18T14:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:59 crc kubenswrapper[4896]: E0218 14:55:59.667315 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3d8f65ef-15b5-4cc3-be66-c8dd6cb956e9\\\",\\\"systemUUID\\\":\\\"d255e81f-4350-4697-bbdf-3f95dec57eda\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:59Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.670478 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.670502 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.670510 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.670523 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.670533 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:59Z","lastTransitionTime":"2026-02-18T14:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:59 crc kubenswrapper[4896]: E0218 14:55:59.681412 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3d8f65ef-15b5-4cc3-be66-c8dd6cb956e9\\\",\\\"systemUUID\\\":\\\"d255e81f-4350-4697-bbdf-3f95dec57eda\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:59Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.685438 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.685479 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.685493 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.685510 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.685521 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:59Z","lastTransitionTime":"2026-02-18T14:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:59 crc kubenswrapper[4896]: E0218 14:55:59.698456 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3d8f65ef-15b5-4cc3-be66-c8dd6cb956e9\\\",\\\"systemUUID\\\":\\\"d255e81f-4350-4697-bbdf-3f95dec57eda\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:59Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.701766 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.701795 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.701804 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.701816 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.701826 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:59Z","lastTransitionTime":"2026-02-18T14:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:59 crc kubenswrapper[4896]: E0218 14:55:59.716932 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3d8f65ef-15b5-4cc3-be66-c8dd6cb956e9\\\",\\\"systemUUID\\\":\\\"d255e81f-4350-4697-bbdf-3f95dec57eda\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:59Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.720896 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.720956 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.720966 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.720985 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.720998 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:59Z","lastTransitionTime":"2026-02-18T14:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:59 crc kubenswrapper[4896]: E0218 14:55:59.749643 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3d8f65ef-15b5-4cc3-be66-c8dd6cb956e9\\\",\\\"systemUUID\\\":\\\"d255e81f-4350-4697-bbdf-3f95dec57eda\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:55:59Z is after 2025-08-24T17:21:41Z" Feb 18 14:55:59 crc kubenswrapper[4896]: E0218 14:55:59.749760 4896 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.754831 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.754874 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.754884 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.754897 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.754906 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:59Z","lastTransitionTime":"2026-02-18T14:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.857071 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.857108 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.857118 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.857132 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.857141 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:59Z","lastTransitionTime":"2026-02-18T14:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.896730 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 16:54:06.975759348 +0000 UTC Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.927809 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:55:59 crc kubenswrapper[4896]: E0218 14:55:59.927929 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.959514 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.959546 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.959557 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.959573 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:55:59 crc kubenswrapper[4896]: I0218 14:55:59.959583 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:55:59Z","lastTransitionTime":"2026-02-18T14:55:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.062156 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.062196 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.062204 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.062217 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.062227 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:00Z","lastTransitionTime":"2026-02-18T14:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.163828 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.163860 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.163869 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.163882 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.163892 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:00Z","lastTransitionTime":"2026-02-18T14:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.266316 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.266350 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.266359 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.266372 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.266382 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:00Z","lastTransitionTime":"2026-02-18T14:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.369340 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.369375 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.369383 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.369396 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.369406 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:00Z","lastTransitionTime":"2026-02-18T14:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.472422 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.472479 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.472503 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.472531 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.472553 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:00Z","lastTransitionTime":"2026-02-18T14:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.575442 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.575481 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.575490 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.575504 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.575514 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:00Z","lastTransitionTime":"2026-02-18T14:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.677560 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.677603 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.677613 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.677627 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.677638 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:00Z","lastTransitionTime":"2026-02-18T14:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.779584 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.779625 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.779646 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.779665 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.779676 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:00Z","lastTransitionTime":"2026-02-18T14:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.881421 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.881460 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.881473 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.881496 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.881508 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:00Z","lastTransitionTime":"2026-02-18T14:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.897828 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 22:00:12.17847755 +0000 UTC Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.927449 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.927521 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:56:00 crc kubenswrapper[4896]: E0218 14:56:00.927554 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.927577 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:56:00 crc kubenswrapper[4896]: E0218 14:56:00.927688 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:56:00 crc kubenswrapper[4896]: E0218 14:56:00.927705 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.983562 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.983667 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.983679 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.983694 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:00 crc kubenswrapper[4896]: I0218 14:56:00.983705 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:00Z","lastTransitionTime":"2026-02-18T14:56:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.086495 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.086538 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.086549 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.086564 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.086574 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:01Z","lastTransitionTime":"2026-02-18T14:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.188618 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.188663 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.188671 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.188684 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.188694 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:01Z","lastTransitionTime":"2026-02-18T14:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.291337 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.291380 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.291390 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.291404 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.291412 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:01Z","lastTransitionTime":"2026-02-18T14:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.394035 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.394087 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.394104 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.394125 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.394140 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:01Z","lastTransitionTime":"2026-02-18T14:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.496677 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.496719 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.496727 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.496743 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.496753 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:01Z","lastTransitionTime":"2026-02-18T14:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.598798 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.598831 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.598841 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.598854 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.598862 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:01Z","lastTransitionTime":"2026-02-18T14:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.701604 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.701643 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.701654 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.701668 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.701680 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:01Z","lastTransitionTime":"2026-02-18T14:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.804029 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.804070 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.804079 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.804095 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.804104 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:01Z","lastTransitionTime":"2026-02-18T14:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.898363 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 14:19:20.954002358 +0000 UTC Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.909606 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.909688 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.909734 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.909754 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.909768 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:01Z","lastTransitionTime":"2026-02-18T14:56:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.928508 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:56:01 crc kubenswrapper[4896]: E0218 14:56:01.928636 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.939229 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-98x74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"229213f7-7a25-4982-a449-0122912ba05c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-98x74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:01Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.950389 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:01Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.960419 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:01Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.973690 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:01Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.988120 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:01Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:01 crc kubenswrapper[4896]: I0218 14:56:01.998337 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:01Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.009171 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32bd695f-cf0b-425a-b764-87b3f678d4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://390187c9005f214d18d43cd6b4f987ccc7d2197f8629997ca4c5cd3ffb49c777\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7021f783fefb51e5fb6b701e58eb9279a0513b9ff37cba435d69e1b7bb1d1a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8j9gb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:02Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.012467 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.012511 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.012523 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.012537 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.012548 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:02Z","lastTransitionTime":"2026-02-18T14:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.021050 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:02Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.036716 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3705371cdc4101eed891047c7e68b95155ea083842e0343ab98763e25ea00f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3705371cdc4101eed891047c7e68b95155ea083842e0343ab98763e25ea00f73\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:55:53Z\\\",\\\"message\\\":\\\"ew object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0218 14:55:53.908904 6577 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nI0218 14:55:53.908922 6577 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0218 14:55:53.908938 6577 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI0218 14:55:53.908947 6577 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}\\\\nI0218 14:55:53.908958 6577 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 2.740953ms\\\\nI0218 14:55:53.908969 6577 services_controller.go:356] Processing sync for service openshift-network-diagnostics/network-check-target for network=default\\\\nF0218 14:55:53.908970 6577 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-w862c_openshift-ovn-kubernetes(449d40af-fe23-4350-b66f-de4ce6614177)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:02Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.048735 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:02Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.062642 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:02Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.074541 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6e1900020612d617118e0cca524ffaafe1fac28997fcf1f73a75fe743b4f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:02Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.083135 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6wprd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8896bb9-c6fe-4111-a49a-9dadf23e6855\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80d2101411bb762554b416ff237ce9ce2455e43c3c5379118322264ffd9dcd94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5jc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6wprd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:02Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.102683 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:02Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.114942 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.115004 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.115019 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.115042 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.115056 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:02Z","lastTransitionTime":"2026-02-18T14:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.119838 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e1cf96a28a07565ec2d56a4ffc32c10e95356b7ebf2399b6b3ba833274a68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:02Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.134375 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"219e676d-1cbc-4503-bef3-7b3c5af4e0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd5e328de9a4e016452b309a56258165a94fdd43c0e26f7444a4a765f3cd11c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec861caaac932e4489277e16914fa252647b155492fec4edeb789ea62e11eac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d26b99304cc56b1e016ba622e79f535b6dfd04c197b698dae1720a1ee44afab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc64cbe4646c598d98e96a7ff130d52867c29c552541b1c7adea6530a582e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bdc64cbe4646c598d98e96a7ff130d52867c29c552541b1c7adea6530a582e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:02Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.148088 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:02Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.159001 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:02Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.217625 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.217683 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.217697 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.217720 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.217735 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:02Z","lastTransitionTime":"2026-02-18T14:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.320052 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.320091 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.320105 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.320126 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.320141 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:02Z","lastTransitionTime":"2026-02-18T14:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.423494 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.423549 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.423557 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.423574 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.423584 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:02Z","lastTransitionTime":"2026-02-18T14:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.526348 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.526392 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.526403 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.526422 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.526433 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:02Z","lastTransitionTime":"2026-02-18T14:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.629376 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.629426 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.629438 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.629454 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.629467 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:02Z","lastTransitionTime":"2026-02-18T14:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.731737 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.731771 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.731782 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.731797 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.731809 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:02Z","lastTransitionTime":"2026-02-18T14:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.833603 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.833646 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.833656 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.833672 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.833683 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:02Z","lastTransitionTime":"2026-02-18T14:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.898492 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 00:09:39.169058519 +0000 UTC Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.926973 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.927053 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:56:02 crc kubenswrapper[4896]: E0218 14:56:02.927078 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.926971 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:56:02 crc kubenswrapper[4896]: E0218 14:56:02.927167 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:56:02 crc kubenswrapper[4896]: E0218 14:56:02.927346 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.935780 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.935825 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.935835 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.935851 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:02 crc kubenswrapper[4896]: I0218 14:56:02.935862 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:02Z","lastTransitionTime":"2026-02-18T14:56:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.038148 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.038189 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.038197 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.038213 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.038222 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:03Z","lastTransitionTime":"2026-02-18T14:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.140423 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.140468 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.140476 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.140497 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.140506 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:03Z","lastTransitionTime":"2026-02-18T14:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.242693 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.242742 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.242752 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.242780 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.242788 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:03Z","lastTransitionTime":"2026-02-18T14:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.344558 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.344599 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.344611 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.344628 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.344640 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:03Z","lastTransitionTime":"2026-02-18T14:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.446332 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.446366 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.446377 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.446394 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.446406 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:03Z","lastTransitionTime":"2026-02-18T14:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.549211 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.549276 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.549285 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.549300 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.549311 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:03Z","lastTransitionTime":"2026-02-18T14:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.651132 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.651181 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.651195 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.651215 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.651269 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:03Z","lastTransitionTime":"2026-02-18T14:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.753356 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.753404 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.753413 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.753426 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.753437 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:03Z","lastTransitionTime":"2026-02-18T14:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.857665 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.857709 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.857718 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.857735 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.857757 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:03Z","lastTransitionTime":"2026-02-18T14:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.900322 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 15:29:03.493775212 +0000 UTC Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.927308 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:56:03 crc kubenswrapper[4896]: E0218 14:56:03.927432 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.959621 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.959654 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.959661 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.959674 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:03 crc kubenswrapper[4896]: I0218 14:56:03.959682 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:03Z","lastTransitionTime":"2026-02-18T14:56:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.062338 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.062382 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.062393 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.062408 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.062418 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:04Z","lastTransitionTime":"2026-02-18T14:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.164931 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.164979 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.164990 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.165011 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.165026 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:04Z","lastTransitionTime":"2026-02-18T14:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.267498 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.267530 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.267538 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.267550 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.267560 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:04Z","lastTransitionTime":"2026-02-18T14:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.369463 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.369510 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.369519 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.369533 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.369542 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:04Z","lastTransitionTime":"2026-02-18T14:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.471832 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.471868 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.471879 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.471896 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.471907 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:04Z","lastTransitionTime":"2026-02-18T14:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.573465 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.573503 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.573513 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.573528 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.573538 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:04Z","lastTransitionTime":"2026-02-18T14:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.675528 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.675615 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.675626 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.675640 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.675648 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:04Z","lastTransitionTime":"2026-02-18T14:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.778301 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.778341 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.778380 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.778400 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.778408 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:04Z","lastTransitionTime":"2026-02-18T14:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.880492 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.880537 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.880548 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.880562 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.880577 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:04Z","lastTransitionTime":"2026-02-18T14:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.900897 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-13 23:13:44.560974168 +0000 UTC Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.927610 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.927703 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:56:04 crc kubenswrapper[4896]: E0218 14:56:04.927770 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:56:04 crc kubenswrapper[4896]: E0218 14:56:04.927776 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.927700 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:56:04 crc kubenswrapper[4896]: E0218 14:56:04.927875 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.983084 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.983118 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.983126 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.983142 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:04 crc kubenswrapper[4896]: I0218 14:56:04.983151 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:04Z","lastTransitionTime":"2026-02-18T14:56:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.086960 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.087008 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.087025 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.087046 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.087057 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:05Z","lastTransitionTime":"2026-02-18T14:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.189470 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.189534 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.189558 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.189586 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.189610 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:05Z","lastTransitionTime":"2026-02-18T14:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.291812 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.291852 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.291863 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.291878 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.291888 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:05Z","lastTransitionTime":"2026-02-18T14:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.393931 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.393967 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.393975 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.393989 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.393997 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:05Z","lastTransitionTime":"2026-02-18T14:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.496638 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.496681 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.496690 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.496705 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.496715 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:05Z","lastTransitionTime":"2026-02-18T14:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.598908 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.598937 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.598946 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.598959 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.598969 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:05Z","lastTransitionTime":"2026-02-18T14:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.701565 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.701612 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.701623 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.701639 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.701664 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:05Z","lastTransitionTime":"2026-02-18T14:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.804215 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.804262 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.804271 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.804283 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.804292 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:05Z","lastTransitionTime":"2026-02-18T14:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.901998 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-18 20:58:06.242746036 +0000 UTC Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.907156 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.907187 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.907202 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.907218 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.907229 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:05Z","lastTransitionTime":"2026-02-18T14:56:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:05 crc kubenswrapper[4896]: I0218 14:56:05.927831 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:56:05 crc kubenswrapper[4896]: E0218 14:56:05.927972 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.009405 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.009436 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.009445 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.009460 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.009470 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:06Z","lastTransitionTime":"2026-02-18T14:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.111622 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.111666 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.111677 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.111706 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.111719 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:06Z","lastTransitionTime":"2026-02-18T14:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.214211 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.214278 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.214288 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.214303 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.214314 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:06Z","lastTransitionTime":"2026-02-18T14:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.316530 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.316561 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.316569 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.316582 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.316591 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:06Z","lastTransitionTime":"2026-02-18T14:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.418588 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.418616 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.418624 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.418637 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.418646 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:06Z","lastTransitionTime":"2026-02-18T14:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.521457 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.521526 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.521539 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.521555 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.521570 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:06Z","lastTransitionTime":"2026-02-18T14:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.623970 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.624031 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.624047 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.624069 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.624082 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:06Z","lastTransitionTime":"2026-02-18T14:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.726709 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.726742 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.726752 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.726766 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.726777 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:06Z","lastTransitionTime":"2026-02-18T14:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.829783 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.829875 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.829890 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.829911 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.829924 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:06Z","lastTransitionTime":"2026-02-18T14:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.902640 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 16:42:44.160421737 +0000 UTC Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.927806 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.927833 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.927846 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:56:06 crc kubenswrapper[4896]: E0218 14:56:06.927957 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:56:06 crc kubenswrapper[4896]: E0218 14:56:06.928089 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:56:06 crc kubenswrapper[4896]: E0218 14:56:06.928302 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.932513 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.932543 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.932553 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.932572 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:06 crc kubenswrapper[4896]: I0218 14:56:06.932583 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:06Z","lastTransitionTime":"2026-02-18T14:56:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.036627 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.036678 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.036691 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.036709 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.036718 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:07Z","lastTransitionTime":"2026-02-18T14:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.139129 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.139220 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.139273 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.139301 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.139320 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:07Z","lastTransitionTime":"2026-02-18T14:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.241302 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.241350 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.241362 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.241378 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.241388 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:07Z","lastTransitionTime":"2026-02-18T14:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.343798 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.343840 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.343849 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.343869 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.343888 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:07Z","lastTransitionTime":"2026-02-18T14:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.447199 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.447283 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.447299 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.447322 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.447339 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:07Z","lastTransitionTime":"2026-02-18T14:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.550582 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.550637 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.550651 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.550676 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.550697 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:07Z","lastTransitionTime":"2026-02-18T14:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.652836 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.652908 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.652919 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.652936 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.652947 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:07Z","lastTransitionTime":"2026-02-18T14:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.755756 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.755802 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.755814 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.755830 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.755842 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:07Z","lastTransitionTime":"2026-02-18T14:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.858631 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.858670 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.858680 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.858694 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.858705 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:07Z","lastTransitionTime":"2026-02-18T14:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.902967 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 20:40:08.983863781 +0000 UTC Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.927430 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:56:07 crc kubenswrapper[4896]: E0218 14:56:07.927590 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.961036 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.961073 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.961082 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.961098 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:07 crc kubenswrapper[4896]: I0218 14:56:07.961107 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:07Z","lastTransitionTime":"2026-02-18T14:56:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.063363 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.063405 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.063417 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.063434 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.063445 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:08Z","lastTransitionTime":"2026-02-18T14:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.166171 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.166211 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.166219 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.166253 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.166264 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:08Z","lastTransitionTime":"2026-02-18T14:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.268302 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.268354 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.268370 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.268392 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.268409 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:08Z","lastTransitionTime":"2026-02-18T14:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.371071 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.371139 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.371152 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.371178 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.371192 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:08Z","lastTransitionTime":"2026-02-18T14:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.474551 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.474633 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.474652 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.474681 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.474703 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:08Z","lastTransitionTime":"2026-02-18T14:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.578582 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.578655 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.578678 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.578715 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.578733 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:08Z","lastTransitionTime":"2026-02-18T14:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.681573 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.681631 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.681641 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.681655 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.681664 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:08Z","lastTransitionTime":"2026-02-18T14:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.784268 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.784319 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.784329 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.784352 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.784363 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:08Z","lastTransitionTime":"2026-02-18T14:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.887117 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.887148 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.887156 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.887169 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.887178 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:08Z","lastTransitionTime":"2026-02-18T14:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.903938 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 09:56:13.904200163 +0000 UTC Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.927364 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.927370 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.927384 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:56:08 crc kubenswrapper[4896]: E0218 14:56:08.927628 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:56:08 crc kubenswrapper[4896]: E0218 14:56:08.927699 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:56:08 crc kubenswrapper[4896]: E0218 14:56:08.927740 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.928368 4896 scope.go:117] "RemoveContainer" containerID="3705371cdc4101eed891047c7e68b95155ea083842e0343ab98763e25ea00f73" Feb 18 14:56:08 crc kubenswrapper[4896]: E0218 14:56:08.928531 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-w862c_openshift-ovn-kubernetes(449d40af-fe23-4350-b66f-de4ce6614177)\"" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" podUID="449d40af-fe23-4350-b66f-de4ce6614177" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.990369 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.990409 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.990418 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.990436 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:08 crc kubenswrapper[4896]: I0218 14:56:08.990448 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:08Z","lastTransitionTime":"2026-02-18T14:56:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.093119 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.093157 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.093166 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.093180 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.093190 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:09Z","lastTransitionTime":"2026-02-18T14:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.195810 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.195847 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.195858 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.195876 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.195888 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:09Z","lastTransitionTime":"2026-02-18T14:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.298381 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.298423 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.298434 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.298455 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.298469 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:09Z","lastTransitionTime":"2026-02-18T14:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.400974 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.401064 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.401108 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.401163 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.401184 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:09Z","lastTransitionTime":"2026-02-18T14:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.504174 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.504221 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.504252 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.504269 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.504281 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:09Z","lastTransitionTime":"2026-02-18T14:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.606205 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.606283 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.606297 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.606313 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.606325 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:09Z","lastTransitionTime":"2026-02-18T14:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.708134 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.708175 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.708184 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.708200 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.708212 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:09Z","lastTransitionTime":"2026-02-18T14:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.810109 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.810155 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.810164 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.810179 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.810188 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:09Z","lastTransitionTime":"2026-02-18T14:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.904279 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 14:50:22.937827106 +0000 UTC Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.912984 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.913041 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.913053 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.913071 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.913085 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:09Z","lastTransitionTime":"2026-02-18T14:56:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:09 crc kubenswrapper[4896]: I0218 14:56:09.927513 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:56:09 crc kubenswrapper[4896]: E0218 14:56:09.927728 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.015219 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.015279 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.015290 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.015308 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.015318 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:10Z","lastTransitionTime":"2026-02-18T14:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.018393 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.018474 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.018491 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.018510 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.018525 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:10Z","lastTransitionTime":"2026-02-18T14:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:10 crc kubenswrapper[4896]: E0218 14:56:10.036592 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3d8f65ef-15b5-4cc3-be66-c8dd6cb956e9\\\",\\\"systemUUID\\\":\\\"d255e81f-4350-4697-bbdf-3f95dec57eda\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:10Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.040427 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.040493 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.040513 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.040536 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.040551 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:10Z","lastTransitionTime":"2026-02-18T14:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:10 crc kubenswrapper[4896]: E0218 14:56:10.055599 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3d8f65ef-15b5-4cc3-be66-c8dd6cb956e9\\\",\\\"systemUUID\\\":\\\"d255e81f-4350-4697-bbdf-3f95dec57eda\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:10Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.060166 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.060204 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.060216 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.060231 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.060262 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:10Z","lastTransitionTime":"2026-02-18T14:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:10 crc kubenswrapper[4896]: E0218 14:56:10.078825 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3d8f65ef-15b5-4cc3-be66-c8dd6cb956e9\\\",\\\"systemUUID\\\":\\\"d255e81f-4350-4697-bbdf-3f95dec57eda\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:10Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.085089 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.085160 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.085172 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.085207 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.085225 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:10Z","lastTransitionTime":"2026-02-18T14:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:10 crc kubenswrapper[4896]: E0218 14:56:10.101420 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3d8f65ef-15b5-4cc3-be66-c8dd6cb956e9\\\",\\\"systemUUID\\\":\\\"d255e81f-4350-4697-bbdf-3f95dec57eda\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:10Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.106692 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.106770 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.106786 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.106803 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.106818 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:10Z","lastTransitionTime":"2026-02-18T14:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:10 crc kubenswrapper[4896]: E0218 14:56:10.119926 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3d8f65ef-15b5-4cc3-be66-c8dd6cb956e9\\\",\\\"systemUUID\\\":\\\"d255e81f-4350-4697-bbdf-3f95dec57eda\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:10Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:10 crc kubenswrapper[4896]: E0218 14:56:10.120060 4896 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.121611 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.121671 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.121685 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.121710 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.121722 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:10Z","lastTransitionTime":"2026-02-18T14:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.224103 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.224131 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.224157 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.224173 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.224183 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:10Z","lastTransitionTime":"2026-02-18T14:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.298674 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/229213f7-7a25-4982-a449-0122912ba05c-metrics-certs\") pod \"network-metrics-daemon-98x74\" (UID: \"229213f7-7a25-4982-a449-0122912ba05c\") " pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:56:10 crc kubenswrapper[4896]: E0218 14:56:10.298848 4896 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 14:56:10 crc kubenswrapper[4896]: E0218 14:56:10.298999 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/229213f7-7a25-4982-a449-0122912ba05c-metrics-certs podName:229213f7-7a25-4982-a449-0122912ba05c nodeName:}" failed. No retries permitted until 2026-02-18 14:56:42.298979733 +0000 UTC m=+101.228313946 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/229213f7-7a25-4982-a449-0122912ba05c-metrics-certs") pod "network-metrics-daemon-98x74" (UID: "229213f7-7a25-4982-a449-0122912ba05c") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.326302 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.326371 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.326381 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.326393 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.326401 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:10Z","lastTransitionTime":"2026-02-18T14:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.428605 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.428680 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.428694 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.428722 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.428736 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:10Z","lastTransitionTime":"2026-02-18T14:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.531040 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.531097 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.531106 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.531122 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.531132 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:10Z","lastTransitionTime":"2026-02-18T14:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.632991 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.633058 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.633070 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.633085 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.633096 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:10Z","lastTransitionTime":"2026-02-18T14:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.735775 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.735813 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.735821 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.735836 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.735844 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:10Z","lastTransitionTime":"2026-02-18T14:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.838209 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.838268 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.838281 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.838296 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.838306 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:10Z","lastTransitionTime":"2026-02-18T14:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.904671 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 09:17:16.798559797 +0000 UTC Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.927938 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:56:10 crc kubenswrapper[4896]: E0218 14:56:10.928064 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.928288 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:56:10 crc kubenswrapper[4896]: E0218 14:56:10.928369 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.928443 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:56:10 crc kubenswrapper[4896]: E0218 14:56:10.928572 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.939956 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.939998 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.940011 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.940027 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:10 crc kubenswrapper[4896]: I0218 14:56:10.940039 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:10Z","lastTransitionTime":"2026-02-18T14:56:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.042525 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.042571 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.042583 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.042600 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.042611 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:11Z","lastTransitionTime":"2026-02-18T14:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.144830 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.144862 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.144872 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.144885 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.144895 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:11Z","lastTransitionTime":"2026-02-18T14:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.247193 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.247250 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.247262 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.247281 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.247290 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:11Z","lastTransitionTime":"2026-02-18T14:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.349019 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.349046 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.349059 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.349077 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.349090 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:11Z","lastTransitionTime":"2026-02-18T14:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.451426 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.451477 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.451486 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.451502 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.451541 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:11Z","lastTransitionTime":"2026-02-18T14:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.553756 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.553793 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.553803 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.553815 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.553825 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:11Z","lastTransitionTime":"2026-02-18T14:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.656720 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.656761 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.656770 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.656787 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.656796 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:11Z","lastTransitionTime":"2026-02-18T14:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.758946 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.758984 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.758998 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.759016 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.759029 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:11Z","lastTransitionTime":"2026-02-18T14:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.862063 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.862111 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.862124 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.862140 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.862150 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:11Z","lastTransitionTime":"2026-02-18T14:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.905617 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 07:17:38.161930486 +0000 UTC Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.927133 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:56:11 crc kubenswrapper[4896]: E0218 14:56:11.927291 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.941303 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:11Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.961355 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3705371cdc4101eed891047c7e68b95155ea083842e0343ab98763e25ea00f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3705371cdc4101eed891047c7e68b95155ea083842e0343ab98763e25ea00f73\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:55:53Z\\\",\\\"message\\\":\\\"ew object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0218 14:55:53.908904 6577 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nI0218 14:55:53.908922 6577 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0218 14:55:53.908938 6577 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI0218 14:55:53.908947 6577 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}\\\\nI0218 14:55:53.908958 6577 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 2.740953ms\\\\nI0218 14:55:53.908969 6577 services_controller.go:356] Processing sync for service openshift-network-diagnostics/network-check-target for network=default\\\\nF0218 14:55:53.908970 6577 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-w862c_openshift-ovn-kubernetes(449d40af-fe23-4350-b66f-de4ce6614177)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:11Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.965289 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.965335 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.965352 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.965369 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.965380 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:11Z","lastTransitionTime":"2026-02-18T14:56:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.975570 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:11Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.990098 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6e1900020612d617118e0cca524ffaafe1fac28997fcf1f73a75fe743b4f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:11Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:11 crc kubenswrapper[4896]: I0218 14:56:11.999898 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6wprd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8896bb9-c6fe-4111-a49a-9dadf23e6855\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80d2101411bb762554b416ff237ce9ce2455e43c3c5379118322264ffd9dcd94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5jc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6wprd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:11Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.009803 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:12Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.022347 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"219e676d-1cbc-4503-bef3-7b3c5af4e0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd5e328de9a4e016452b309a56258165a94fdd43c0e26f7444a4a765f3cd11c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec861caaac932e4489277e16914fa252647b155492fec4edeb789ea62e11eac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d26b99304cc56b1e016ba622e79f535b6dfd04c197b698dae1720a1ee44afab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc64cbe4646c598d98e96a7ff130d52867c29c552541b1c7adea6530a582e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bdc64cbe4646c598d98e96a7ff130d52867c29c552541b1c7adea6530a582e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:12Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.033531 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:12Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.042755 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:12Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.059989 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:12Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.066515 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.066561 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.066570 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.066584 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.066592 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:12Z","lastTransitionTime":"2026-02-18T14:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.071946 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e1cf96a28a07565ec2d56a4ffc32c10e95356b7ebf2399b6b3ba833274a68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:12Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.083994 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:12Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.096200 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:12Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.104290 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:12Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.114127 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32bd695f-cf0b-425a-b764-87b3f678d4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://390187c9005f214d18d43cd6b4f987ccc7d2197f8629997ca4c5cd3ffb49c777\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7021f783fefb51e5fb6b701e58eb9279a0513b9ff37cba435d69e1b7bb1d1a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8j9gb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:12Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.122992 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-98x74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"229213f7-7a25-4982-a449-0122912ba05c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-98x74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:12Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.133438 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:12Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.142258 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:12Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.169601 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.169630 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.169642 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.169659 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.169670 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:12Z","lastTransitionTime":"2026-02-18T14:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.271218 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.271460 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.271555 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.271630 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.271690 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:12Z","lastTransitionTime":"2026-02-18T14:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.335854 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fmg2f_c5ba4981-ee97-4d71-b48f-b5d12bd5911c/kube-multus/0.log" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.335908 4896 generic.go:334] "Generic (PLEG): container finished" podID="c5ba4981-ee97-4d71-b48f-b5d12bd5911c" containerID="434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39" exitCode=1 Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.335940 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fmg2f" event={"ID":"c5ba4981-ee97-4d71-b48f-b5d12bd5911c","Type":"ContainerDied","Data":"434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39"} Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.336309 4896 scope.go:117] "RemoveContainer" containerID="434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.348436 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:12Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.359582 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:12Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.372700 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6e1900020612d617118e0cca524ffaafe1fac28997fcf1f73a75fe743b4f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:12Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.374683 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.374720 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.374730 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.374746 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.374757 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:12Z","lastTransitionTime":"2026-02-18T14:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.381182 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6wprd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8896bb9-c6fe-4111-a49a-9dadf23e6855\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80d2101411bb762554b416ff237ce9ce2455e43c3c5379118322264ffd9dcd94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5jc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6wprd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:12Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.400326 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:12Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.412133 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e1cf96a28a07565ec2d56a4ffc32c10e95356b7ebf2399b6b3ba833274a68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:12Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.422837 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"219e676d-1cbc-4503-bef3-7b3c5af4e0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd5e328de9a4e016452b309a56258165a94fdd43c0e26f7444a4a765f3cd11c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec861caaac932e4489277e16914fa252647b155492fec4edeb789ea62e11eac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d26b99304cc56b1e016ba622e79f535b6dfd04c197b698dae1720a1ee44afab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc64cbe4646c598d98e96a7ff130d52867c29c552541b1c7adea6530a582e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bdc64cbe4646c598d98e96a7ff130d52867c29c552541b1c7adea6530a582e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:12Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.435230 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:12Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.445841 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:12Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.456736 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:12Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.468927 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:12Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.476510 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.476539 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.476550 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.476565 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.476577 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:12Z","lastTransitionTime":"2026-02-18T14:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.480347 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:12Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.491737 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:12Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.510480 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:12Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.519345 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32bd695f-cf0b-425a-b764-87b3f678d4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://390187c9005f214d18d43cd6b4f987ccc7d2197f8629997ca4c5cd3ffb49c777\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7021f783fefb51e5fb6b701e58eb9279a0513b9ff37cba435d69e1b7bb1d1a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8j9gb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:12Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.527967 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-98x74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"229213f7-7a25-4982-a449-0122912ba05c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-98x74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:12Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.540433 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:56:11Z\\\",\\\"message\\\":\\\"2026-02-18T14:55:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_535ee55e-8811-41f2-8919-a177df470c76\\\\n2026-02-18T14:55:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_535ee55e-8811-41f2-8919-a177df470c76 to /host/opt/cni/bin/\\\\n2026-02-18T14:55:26Z [verbose] multus-daemon started\\\\n2026-02-18T14:55:26Z [verbose] Readiness Indicator file check\\\\n2026-02-18T14:56:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:12Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.558802 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3705371cdc4101eed891047c7e68b95155ea083842e0343ab98763e25ea00f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3705371cdc4101eed891047c7e68b95155ea083842e0343ab98763e25ea00f73\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:55:53Z\\\",\\\"message\\\":\\\"ew object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0218 14:55:53.908904 6577 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nI0218 14:55:53.908922 6577 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0218 14:55:53.908938 6577 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI0218 14:55:53.908947 6577 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}\\\\nI0218 14:55:53.908958 6577 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 2.740953ms\\\\nI0218 14:55:53.908969 6577 services_controller.go:356] Processing sync for service openshift-network-diagnostics/network-check-target for network=default\\\\nF0218 14:55:53.908970 6577 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-w862c_openshift-ovn-kubernetes(449d40af-fe23-4350-b66f-de4ce6614177)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:12Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.578472 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.578508 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.578519 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.578533 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.578543 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:12Z","lastTransitionTime":"2026-02-18T14:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.680779 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.680824 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.680837 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.680853 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.680863 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:12Z","lastTransitionTime":"2026-02-18T14:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.783128 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.783161 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.783171 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.783186 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.783197 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:12Z","lastTransitionTime":"2026-02-18T14:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.884844 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.884875 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.884886 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.884945 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.884958 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:12Z","lastTransitionTime":"2026-02-18T14:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.906054 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 13:33:35.133820248 +0000 UTC Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.927522 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.927551 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.927609 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:56:12 crc kubenswrapper[4896]: E0218 14:56:12.927615 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:56:12 crc kubenswrapper[4896]: E0218 14:56:12.927710 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:56:12 crc kubenswrapper[4896]: E0218 14:56:12.927762 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.987298 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.987337 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.987345 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.987359 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:12 crc kubenswrapper[4896]: I0218 14:56:12.987368 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:12Z","lastTransitionTime":"2026-02-18T14:56:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.089930 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.089962 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.089972 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.089985 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.089994 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:13Z","lastTransitionTime":"2026-02-18T14:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.192609 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.192645 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.192655 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.192668 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.192680 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:13Z","lastTransitionTime":"2026-02-18T14:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.294947 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.294987 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.294996 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.295007 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.295015 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:13Z","lastTransitionTime":"2026-02-18T14:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.340535 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fmg2f_c5ba4981-ee97-4d71-b48f-b5d12bd5911c/kube-multus/0.log" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.340593 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fmg2f" event={"ID":"c5ba4981-ee97-4d71-b48f-b5d12bd5911c","Type":"ContainerStarted","Data":"54f94d214d96548b9d5a424bc162f56ffecc188391e16ee0c3b4009d55ca0472"} Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.351639 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6wprd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8896bb9-c6fe-4111-a49a-9dadf23e6855\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80d2101411bb762554b416ff237ce9ce2455e43c3c5379118322264ffd9dcd94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5jc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6wprd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:13Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.363763 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:13Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.375977 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:13Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.390186 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6e1900020612d617118e0cca524ffaafe1fac28997fcf1f73a75fe743b4f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:13Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.397791 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.397821 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.397832 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.397848 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.397860 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:13Z","lastTransitionTime":"2026-02-18T14:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.403927 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:13Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.428066 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:13Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.442513 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e1cf96a28a07565ec2d56a4ffc32c10e95356b7ebf2399b6b3ba833274a68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:13Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.452631 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"219e676d-1cbc-4503-bef3-7b3c5af4e0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd5e328de9a4e016452b309a56258165a94fdd43c0e26f7444a4a765f3cd11c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec861caaac932e4489277e16914fa252647b155492fec4edeb789ea62e11eac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d26b99304cc56b1e016ba622e79f535b6dfd04c197b698dae1720a1ee44afab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc64cbe4646c598d98e96a7ff130d52867c29c552541b1c7adea6530a582e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bdc64cbe4646c598d98e96a7ff130d52867c29c552541b1c7adea6530a582e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:13Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.466748 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:13Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.477549 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:13Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.487632 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32bd695f-cf0b-425a-b764-87b3f678d4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://390187c9005f214d18d43cd6b4f987ccc7d2197f8629997ca4c5cd3ffb49c777\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7021f783fefb51e5fb6b701e58eb9279a0513b9ff37cba435d69e1b7bb1d1a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8j9gb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:13Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.497707 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-98x74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"229213f7-7a25-4982-a449-0122912ba05c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-98x74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:13Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.500419 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.500447 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.500457 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.500473 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.500483 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:13Z","lastTransitionTime":"2026-02-18T14:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.509661 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:13Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.519309 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:13Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.532111 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:13Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.544275 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:13Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.557119 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54f94d214d96548b9d5a424bc162f56ffecc188391e16ee0c3b4009d55ca0472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:56:11Z\\\",\\\"message\\\":\\\"2026-02-18T14:55:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_535ee55e-8811-41f2-8919-a177df470c76\\\\n2026-02-18T14:55:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_535ee55e-8811-41f2-8919-a177df470c76 to /host/opt/cni/bin/\\\\n2026-02-18T14:55:26Z [verbose] multus-daemon started\\\\n2026-02-18T14:55:26Z [verbose] Readiness Indicator file check\\\\n2026-02-18T14:56:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:13Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.576360 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3705371cdc4101eed891047c7e68b95155ea083842e0343ab98763e25ea00f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3705371cdc4101eed891047c7e68b95155ea083842e0343ab98763e25ea00f73\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:55:53Z\\\",\\\"message\\\":\\\"ew object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0218 14:55:53.908904 6577 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nI0218 14:55:53.908922 6577 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0218 14:55:53.908938 6577 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI0218 14:55:53.908947 6577 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}\\\\nI0218 14:55:53.908958 6577 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 2.740953ms\\\\nI0218 14:55:53.908969 6577 services_controller.go:356] Processing sync for service openshift-network-diagnostics/network-check-target for network=default\\\\nF0218 14:55:53.908970 6577 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-w862c_openshift-ovn-kubernetes(449d40af-fe23-4350-b66f-de4ce6614177)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:13Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.602523 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.602562 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.602572 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.602586 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.602594 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:13Z","lastTransitionTime":"2026-02-18T14:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.704846 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.705132 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.705220 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.705331 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.705423 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:13Z","lastTransitionTime":"2026-02-18T14:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.808094 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.808144 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.808159 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.808177 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.808193 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:13Z","lastTransitionTime":"2026-02-18T14:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.906682 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 05:00:32.855894317 +0000 UTC Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.910186 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.910247 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.910257 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.910272 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.910281 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:13Z","lastTransitionTime":"2026-02-18T14:56:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:13 crc kubenswrapper[4896]: I0218 14:56:13.927902 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:56:13 crc kubenswrapper[4896]: E0218 14:56:13.928212 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.012787 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.012824 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.012833 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.012845 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.012853 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:14Z","lastTransitionTime":"2026-02-18T14:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.114734 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.114771 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.114779 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.114792 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.114801 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:14Z","lastTransitionTime":"2026-02-18T14:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.217202 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.217251 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.217260 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.217274 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.217283 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:14Z","lastTransitionTime":"2026-02-18T14:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.319911 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.319956 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.319979 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.319998 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.320010 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:14Z","lastTransitionTime":"2026-02-18T14:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.422543 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.422580 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.422593 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.422609 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.422621 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:14Z","lastTransitionTime":"2026-02-18T14:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.525200 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.525249 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.525258 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.525271 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.525280 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:14Z","lastTransitionTime":"2026-02-18T14:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.627603 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.627668 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.627690 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.627713 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.627731 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:14Z","lastTransitionTime":"2026-02-18T14:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.730410 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.730451 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.730462 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.730477 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.730488 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:14Z","lastTransitionTime":"2026-02-18T14:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.832885 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.832938 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.832953 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.832971 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.832982 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:14Z","lastTransitionTime":"2026-02-18T14:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.907778 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 11:09:59.467890444 +0000 UTC Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.927109 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.927183 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.927109 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:56:14 crc kubenswrapper[4896]: E0218 14:56:14.927217 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:56:14 crc kubenswrapper[4896]: E0218 14:56:14.927338 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:56:14 crc kubenswrapper[4896]: E0218 14:56:14.927421 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.935317 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.935354 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.935369 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.935385 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:14 crc kubenswrapper[4896]: I0218 14:56:14.935395 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:14Z","lastTransitionTime":"2026-02-18T14:56:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.037926 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.037973 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.037984 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.038001 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.038014 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:15Z","lastTransitionTime":"2026-02-18T14:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.140863 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.140920 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.140933 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.140950 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.140961 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:15Z","lastTransitionTime":"2026-02-18T14:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.244168 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.244220 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.244254 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.244276 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.244298 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:15Z","lastTransitionTime":"2026-02-18T14:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.346839 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.346902 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.346926 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.346958 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.346981 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:15Z","lastTransitionTime":"2026-02-18T14:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.449172 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.449209 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.449217 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.449232 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.449253 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:15Z","lastTransitionTime":"2026-02-18T14:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.550991 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.551038 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.551061 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.551078 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.551090 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:15Z","lastTransitionTime":"2026-02-18T14:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.652949 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.652975 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.652985 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.652999 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.653010 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:15Z","lastTransitionTime":"2026-02-18T14:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.755165 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.755204 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.755223 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.755270 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.755281 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:15Z","lastTransitionTime":"2026-02-18T14:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.857461 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.857498 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.857510 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.857524 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.857534 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:15Z","lastTransitionTime":"2026-02-18T14:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.908173 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 21:25:34.72402074 +0000 UTC Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.927715 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:56:15 crc kubenswrapper[4896]: E0218 14:56:15.927851 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.959919 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.959946 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.959954 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.959982 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:15 crc kubenswrapper[4896]: I0218 14:56:15.959991 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:15Z","lastTransitionTime":"2026-02-18T14:56:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.062636 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.062695 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.062707 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.062725 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.062737 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:16Z","lastTransitionTime":"2026-02-18T14:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.165878 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.165918 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.165932 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.165949 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.165961 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:16Z","lastTransitionTime":"2026-02-18T14:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.268316 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.268382 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.268390 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.268409 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.268421 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:16Z","lastTransitionTime":"2026-02-18T14:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.369843 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.369881 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.369892 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.369907 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.369918 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:16Z","lastTransitionTime":"2026-02-18T14:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.472617 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.472690 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.472708 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.472730 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.472747 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:16Z","lastTransitionTime":"2026-02-18T14:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.575695 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.575770 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.575793 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.575820 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.575836 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:16Z","lastTransitionTime":"2026-02-18T14:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.678530 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.678585 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.678602 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.678627 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.678643 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:16Z","lastTransitionTime":"2026-02-18T14:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.781894 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.781953 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.781970 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.781993 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.782010 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:16Z","lastTransitionTime":"2026-02-18T14:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.884115 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.884148 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.884158 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.884174 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.884184 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:16Z","lastTransitionTime":"2026-02-18T14:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.908808 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 21:56:18.07032091 +0000 UTC Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.927809 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:56:16 crc kubenswrapper[4896]: E0218 14:56:16.927985 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.928405 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:56:16 crc kubenswrapper[4896]: E0218 14:56:16.928529 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.928753 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:56:16 crc kubenswrapper[4896]: E0218 14:56:16.928853 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.986363 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.986391 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.986399 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.986410 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:16 crc kubenswrapper[4896]: I0218 14:56:16.986418 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:16Z","lastTransitionTime":"2026-02-18T14:56:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.088592 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.088627 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.088638 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.088655 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.088667 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:17Z","lastTransitionTime":"2026-02-18T14:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.191357 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.191394 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.191405 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.191420 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.191431 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:17Z","lastTransitionTime":"2026-02-18T14:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.293922 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.294008 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.294026 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.294043 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.294055 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:17Z","lastTransitionTime":"2026-02-18T14:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.396337 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.396375 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.396387 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.396403 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.396414 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:17Z","lastTransitionTime":"2026-02-18T14:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.498718 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.498755 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.498763 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.498779 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.498788 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:17Z","lastTransitionTime":"2026-02-18T14:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.601329 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.601490 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.601513 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.601573 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.601594 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:17Z","lastTransitionTime":"2026-02-18T14:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.704215 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.704367 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.704389 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.704413 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.704433 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:17Z","lastTransitionTime":"2026-02-18T14:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.807217 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.807290 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.807305 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.807322 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.807333 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:17Z","lastTransitionTime":"2026-02-18T14:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.908990 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 06:08:18.78653133 +0000 UTC Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.909916 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.909979 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.909991 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.910005 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.910016 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:17Z","lastTransitionTime":"2026-02-18T14:56:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:17 crc kubenswrapper[4896]: I0218 14:56:17.927598 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:56:17 crc kubenswrapper[4896]: E0218 14:56:17.927815 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.012210 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.012267 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.012279 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.012297 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.012309 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:18Z","lastTransitionTime":"2026-02-18T14:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.114096 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.114131 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.114141 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.114157 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.114166 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:18Z","lastTransitionTime":"2026-02-18T14:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.216414 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.216446 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.216453 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.216466 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.216474 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:18Z","lastTransitionTime":"2026-02-18T14:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.319153 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.319181 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.319189 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.319201 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.319208 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:18Z","lastTransitionTime":"2026-02-18T14:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.421140 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.421190 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.421201 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.421220 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.421256 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:18Z","lastTransitionTime":"2026-02-18T14:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.523391 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.523448 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.523464 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.523489 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.523508 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:18Z","lastTransitionTime":"2026-02-18T14:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.625852 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.625879 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.625888 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.625900 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.625909 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:18Z","lastTransitionTime":"2026-02-18T14:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.728052 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.728082 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.728090 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.728102 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.728111 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:18Z","lastTransitionTime":"2026-02-18T14:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.829983 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.830021 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.830033 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.830049 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.830058 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:18Z","lastTransitionTime":"2026-02-18T14:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.909784 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 14:01:50.334593273 +0000 UTC Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.927376 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.927410 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.927449 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:56:18 crc kubenswrapper[4896]: E0218 14:56:18.927541 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:56:18 crc kubenswrapper[4896]: E0218 14:56:18.927632 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:56:18 crc kubenswrapper[4896]: E0218 14:56:18.927899 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.932119 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.932158 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.932170 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.932186 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:18 crc kubenswrapper[4896]: I0218 14:56:18.932198 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:18Z","lastTransitionTime":"2026-02-18T14:56:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.035019 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.035057 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.035069 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.035087 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.035098 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:19Z","lastTransitionTime":"2026-02-18T14:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.138160 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.138208 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.138219 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.138303 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.138320 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:19Z","lastTransitionTime":"2026-02-18T14:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.240694 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.240727 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.240734 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.240747 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.240756 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:19Z","lastTransitionTime":"2026-02-18T14:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.343550 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.343595 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.343604 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.343618 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.343627 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:19Z","lastTransitionTime":"2026-02-18T14:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.445771 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.445816 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.445828 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.445841 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.445851 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:19Z","lastTransitionTime":"2026-02-18T14:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.547883 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.547908 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.547951 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.547963 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.547973 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:19Z","lastTransitionTime":"2026-02-18T14:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.650801 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.650865 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.650886 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.650916 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.650936 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:19Z","lastTransitionTime":"2026-02-18T14:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.753189 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.753222 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.753249 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.753272 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.753283 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:19Z","lastTransitionTime":"2026-02-18T14:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.855044 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.855072 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.855080 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.855091 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.855100 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:19Z","lastTransitionTime":"2026-02-18T14:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.910095 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-06 19:19:26.140565385 +0000 UTC Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.929555 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:56:19 crc kubenswrapper[4896]: E0218 14:56:19.929678 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.957351 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.957396 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.957404 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.957418 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:19 crc kubenswrapper[4896]: I0218 14:56:19.957427 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:19Z","lastTransitionTime":"2026-02-18T14:56:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.059956 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.060005 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.060018 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.060036 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.060050 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:20Z","lastTransitionTime":"2026-02-18T14:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.162039 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.162079 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.162087 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.162101 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.162109 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:20Z","lastTransitionTime":"2026-02-18T14:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.264520 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.264569 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.264578 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.264592 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.264602 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:20Z","lastTransitionTime":"2026-02-18T14:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.317348 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.317402 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.317413 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.317430 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.317441 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:20Z","lastTransitionTime":"2026-02-18T14:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:20 crc kubenswrapper[4896]: E0218 14:56:20.333228 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3d8f65ef-15b5-4cc3-be66-c8dd6cb956e9\\\",\\\"systemUUID\\\":\\\"d255e81f-4350-4697-bbdf-3f95dec57eda\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:20Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.336617 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.336664 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.336674 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.336688 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.336697 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:20Z","lastTransitionTime":"2026-02-18T14:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:20 crc kubenswrapper[4896]: E0218 14:56:20.349002 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3d8f65ef-15b5-4cc3-be66-c8dd6cb956e9\\\",\\\"systemUUID\\\":\\\"d255e81f-4350-4697-bbdf-3f95dec57eda\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:20Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.352558 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.352581 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.352591 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.352606 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.352618 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:20Z","lastTransitionTime":"2026-02-18T14:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:20 crc kubenswrapper[4896]: E0218 14:56:20.363632 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3d8f65ef-15b5-4cc3-be66-c8dd6cb956e9\\\",\\\"systemUUID\\\":\\\"d255e81f-4350-4697-bbdf-3f95dec57eda\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:20Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.368269 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.368442 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.368531 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.368624 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.368853 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:20Z","lastTransitionTime":"2026-02-18T14:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:20 crc kubenswrapper[4896]: E0218 14:56:20.385323 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3d8f65ef-15b5-4cc3-be66-c8dd6cb956e9\\\",\\\"systemUUID\\\":\\\"d255e81f-4350-4697-bbdf-3f95dec57eda\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:20Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.388462 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.388649 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.388760 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.388875 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.388980 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:20Z","lastTransitionTime":"2026-02-18T14:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:20 crc kubenswrapper[4896]: E0218 14:56:20.401764 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"3d8f65ef-15b5-4cc3-be66-c8dd6cb956e9\\\",\\\"systemUUID\\\":\\\"d255e81f-4350-4697-bbdf-3f95dec57eda\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:20Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:20 crc kubenswrapper[4896]: E0218 14:56:20.401908 4896 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.403263 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.403324 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.403334 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.403346 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.403355 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:20Z","lastTransitionTime":"2026-02-18T14:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.504763 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.504796 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.504806 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.504818 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.504843 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:20Z","lastTransitionTime":"2026-02-18T14:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.606445 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.606487 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.606497 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.606512 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.606523 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:20Z","lastTransitionTime":"2026-02-18T14:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.708314 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.708364 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.708379 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.708400 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.708418 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:20Z","lastTransitionTime":"2026-02-18T14:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.814076 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.814107 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.814140 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.814156 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.814164 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:20Z","lastTransitionTime":"2026-02-18T14:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.910383 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 16:59:02.667246836 +0000 UTC Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.915927 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.915960 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.915975 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.915995 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.916010 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:20Z","lastTransitionTime":"2026-02-18T14:56:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.927509 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.927530 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:56:20 crc kubenswrapper[4896]: I0218 14:56:20.927513 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:56:20 crc kubenswrapper[4896]: E0218 14:56:20.927619 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:56:20 crc kubenswrapper[4896]: E0218 14:56:20.927698 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:56:20 crc kubenswrapper[4896]: E0218 14:56:20.927828 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.019158 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.019204 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.019218 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.019259 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.019271 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:21Z","lastTransitionTime":"2026-02-18T14:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.122766 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.122869 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.122891 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.122916 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.122936 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:21Z","lastTransitionTime":"2026-02-18T14:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.224984 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.225026 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.225036 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.225052 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.225065 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:21Z","lastTransitionTime":"2026-02-18T14:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.327770 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.327826 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.327838 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.327853 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.327866 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:21Z","lastTransitionTime":"2026-02-18T14:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.430148 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.430208 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.430226 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.430284 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.430308 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:21Z","lastTransitionTime":"2026-02-18T14:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.532701 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.532805 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.532826 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.532891 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.532910 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:21Z","lastTransitionTime":"2026-02-18T14:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.636143 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.636199 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.636222 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.636334 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.636371 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:21Z","lastTransitionTime":"2026-02-18T14:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.739095 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.739155 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.739172 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.739196 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.739216 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:21Z","lastTransitionTime":"2026-02-18T14:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.841565 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.841615 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.841626 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.841642 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.841652 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:21Z","lastTransitionTime":"2026-02-18T14:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.910865 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-31 09:18:48.172059094 +0000 UTC Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.927584 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:56:21 crc kubenswrapper[4896]: E0218 14:56:21.927739 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.943941 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.943983 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.943994 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.944010 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.944022 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:21Z","lastTransitionTime":"2026-02-18T14:56:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.944805 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-98x74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"229213f7-7a25-4982-a449-0122912ba05c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-98x74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:21Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.964625 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:21Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.981494 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:21Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:21 crc kubenswrapper[4896]: I0218 14:56:21.998952 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:21Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.019147 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:22Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.035453 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:22Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.046537 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.046561 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.046569 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.046582 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.046591 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:22Z","lastTransitionTime":"2026-02-18T14:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.053376 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32bd695f-cf0b-425a-b764-87b3f678d4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://390187c9005f214d18d43cd6b4f987ccc7d2197f8629997ca4c5cd3ffb49c777\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7021f783fefb51e5fb6b701e58eb9279a0513b9ff37cba435d69e1b7bb1d1a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8j9gb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:22Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.073437 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54f94d214d96548b9d5a424bc162f56ffecc188391e16ee0c3b4009d55ca0472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:56:11Z\\\",\\\"message\\\":\\\"2026-02-18T14:55:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_535ee55e-8811-41f2-8919-a177df470c76\\\\n2026-02-18T14:55:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_535ee55e-8811-41f2-8919-a177df470c76 to /host/opt/cni/bin/\\\\n2026-02-18T14:55:26Z [verbose] multus-daemon started\\\\n2026-02-18T14:55:26Z [verbose] Readiness Indicator file check\\\\n2026-02-18T14:56:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:22Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.101826 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3705371cdc4101eed891047c7e68b95155ea083842e0343ab98763e25ea00f73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3705371cdc4101eed891047c7e68b95155ea083842e0343ab98763e25ea00f73\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:55:53Z\\\",\\\"message\\\":\\\"ew object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0218 14:55:53.908904 6577 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nI0218 14:55:53.908922 6577 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0218 14:55:53.908938 6577 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI0218 14:55:53.908947 6577 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}\\\\nI0218 14:55:53.908958 6577 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 2.740953ms\\\\nI0218 14:55:53.908969 6577 services_controller.go:356] Processing sync for service openshift-network-diagnostics/network-check-target for network=default\\\\nF0218 14:55:53.908970 6577 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-w862c_openshift-ovn-kubernetes(449d40af-fe23-4350-b66f-de4ce6614177)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:22Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.117521 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:22Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.130632 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:22Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.145645 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6e1900020612d617118e0cca524ffaafe1fac28997fcf1f73a75fe743b4f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:22Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.149343 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.149390 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.149404 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.149425 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.149441 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:22Z","lastTransitionTime":"2026-02-18T14:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.157447 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6wprd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8896bb9-c6fe-4111-a49a-9dadf23e6855\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80d2101411bb762554b416ff237ce9ce2455e43c3c5379118322264ffd9dcd94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5jc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6wprd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:22Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.180271 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:22Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.194207 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e1cf96a28a07565ec2d56a4ffc32c10e95356b7ebf2399b6b3ba833274a68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:22Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.204515 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"219e676d-1cbc-4503-bef3-7b3c5af4e0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd5e328de9a4e016452b309a56258165a94fdd43c0e26f7444a4a765f3cd11c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec861caaac932e4489277e16914fa252647b155492fec4edeb789ea62e11eac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d26b99304cc56b1e016ba622e79f535b6dfd04c197b698dae1720a1ee44afab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc64cbe4646c598d98e96a7ff130d52867c29c552541b1c7adea6530a582e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bdc64cbe4646c598d98e96a7ff130d52867c29c552541b1c7adea6530a582e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:22Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.215746 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:22Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.225750 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:22Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.251821 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.251848 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.251856 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.251870 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.251878 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:22Z","lastTransitionTime":"2026-02-18T14:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.354675 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.354713 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.354726 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.354741 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.354755 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:22Z","lastTransitionTime":"2026-02-18T14:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.457640 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.457683 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.457694 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.457711 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.457723 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:22Z","lastTransitionTime":"2026-02-18T14:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.559650 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.559736 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.559748 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.559765 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.559776 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:22Z","lastTransitionTime":"2026-02-18T14:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.662698 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.662772 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.662790 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.663376 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.663409 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:22Z","lastTransitionTime":"2026-02-18T14:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.766075 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.766117 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.766127 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.766141 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.766150 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:22Z","lastTransitionTime":"2026-02-18T14:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.868685 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.868734 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.868747 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.868764 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.868775 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:22Z","lastTransitionTime":"2026-02-18T14:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.911107 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 02:58:09.075366175 +0000 UTC Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.927576 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.927592 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.927645 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:56:22 crc kubenswrapper[4896]: E0218 14:56:22.927713 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:56:22 crc kubenswrapper[4896]: E0218 14:56:22.927812 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:56:22 crc kubenswrapper[4896]: E0218 14:56:22.928355 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.928702 4896 scope.go:117] "RemoveContainer" containerID="3705371cdc4101eed891047c7e68b95155ea083842e0343ab98763e25ea00f73" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.971578 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.971935 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.971958 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.971985 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:22 crc kubenswrapper[4896]: I0218 14:56:22.972008 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:22Z","lastTransitionTime":"2026-02-18T14:56:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.074090 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.074346 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.074425 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.074546 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.074624 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:23Z","lastTransitionTime":"2026-02-18T14:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.176384 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.176442 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.176459 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.176479 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.176494 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:23Z","lastTransitionTime":"2026-02-18T14:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.279168 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.279196 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.279320 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.279336 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.279347 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:23Z","lastTransitionTime":"2026-02-18T14:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.371077 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w862c_449d40af-fe23-4350-b66f-de4ce6614177/ovnkube-controller/2.log" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.373543 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" event={"ID":"449d40af-fe23-4350-b66f-de4ce6614177","Type":"ContainerStarted","Data":"670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407"} Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.373980 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.384315 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.384344 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.384354 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.384369 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.384378 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:23Z","lastTransitionTime":"2026-02-18T14:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.385101 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32bd695f-cf0b-425a-b764-87b3f678d4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://390187c9005f214d18d43cd6b4f987ccc7d2197f8629997ca4c5cd3ffb49c777\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7021f783fefb51e5fb6b701e58eb9279a0513b9ff37cba435d69e1b7bb1d1a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8j9gb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:23Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.405203 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-98x74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"229213f7-7a25-4982-a449-0122912ba05c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-98x74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:23Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.477989 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:23Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.486673 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.486712 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.486721 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.486735 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.486747 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:23Z","lastTransitionTime":"2026-02-18T14:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.499925 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:23Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.512988 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:23Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.523852 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:23Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.532786 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:23Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.542377 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54f94d214d96548b9d5a424bc162f56ffecc188391e16ee0c3b4009d55ca0472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:56:11Z\\\",\\\"message\\\":\\\"2026-02-18T14:55:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_535ee55e-8811-41f2-8919-a177df470c76\\\\n2026-02-18T14:55:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_535ee55e-8811-41f2-8919-a177df470c76 to /host/opt/cni/bin/\\\\n2026-02-18T14:55:26Z [verbose] multus-daemon started\\\\n2026-02-18T14:55:26Z [verbose] Readiness Indicator file check\\\\n2026-02-18T14:56:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:23Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.559863 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3705371cdc4101eed891047c7e68b95155ea083842e0343ab98763e25ea00f73\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:55:53Z\\\",\\\"message\\\":\\\"ew object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0218 14:55:53.908904 6577 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nI0218 14:55:53.908922 6577 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0218 14:55:53.908938 6577 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI0218 14:55:53.908947 6577 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}\\\\nI0218 14:55:53.908958 6577 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 2.740953ms\\\\nI0218 14:55:53.908969 6577 services_controller.go:356] Processing sync for service openshift-network-diagnostics/network-check-target for network=default\\\\nF0218 14:55:53.908970 6577 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:23Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.570612 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:23Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.586174 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:23Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.588729 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.588768 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.588781 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.588798 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.588809 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:23Z","lastTransitionTime":"2026-02-18T14:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.605001 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6e1900020612d617118e0cca524ffaafe1fac28997fcf1f73a75fe743b4f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:23Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.614472 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6wprd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8896bb9-c6fe-4111-a49a-9dadf23e6855\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80d2101411bb762554b416ff237ce9ce2455e43c3c5379118322264ffd9dcd94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5jc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6wprd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:23Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.630911 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:23Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.646998 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e1cf96a28a07565ec2d56a4ffc32c10e95356b7ebf2399b6b3ba833274a68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:23Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.657296 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"219e676d-1cbc-4503-bef3-7b3c5af4e0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd5e328de9a4e016452b309a56258165a94fdd43c0e26f7444a4a765f3cd11c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec861caaac932e4489277e16914fa252647b155492fec4edeb789ea62e11eac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d26b99304cc56b1e016ba622e79f535b6dfd04c197b698dae1720a1ee44afab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc64cbe4646c598d98e96a7ff130d52867c29c552541b1c7adea6530a582e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bdc64cbe4646c598d98e96a7ff130d52867c29c552541b1c7adea6530a582e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:23Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.668858 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:23Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.681944 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:23Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.690403 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.690437 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.690445 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.690459 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.690469 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:23Z","lastTransitionTime":"2026-02-18T14:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.792124 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.792159 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.792169 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.792183 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.792194 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:23Z","lastTransitionTime":"2026-02-18T14:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.894460 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.894501 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.894511 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.894526 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.894536 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:23Z","lastTransitionTime":"2026-02-18T14:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.911837 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 13:18:35.437742261 +0000 UTC Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.927223 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:56:23 crc kubenswrapper[4896]: E0218 14:56:23.927366 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.996916 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.996984 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.996999 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.997020 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:23 crc kubenswrapper[4896]: I0218 14:56:23.997081 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:23Z","lastTransitionTime":"2026-02-18T14:56:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.099884 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.099911 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.099919 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.099930 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.099938 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:24Z","lastTransitionTime":"2026-02-18T14:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.202029 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.202057 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.202081 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.202092 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.202102 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:24Z","lastTransitionTime":"2026-02-18T14:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.304040 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.304072 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.304080 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.304093 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.304102 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:24Z","lastTransitionTime":"2026-02-18T14:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.378655 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w862c_449d40af-fe23-4350-b66f-de4ce6614177/ovnkube-controller/3.log" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.379346 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w862c_449d40af-fe23-4350-b66f-de4ce6614177/ovnkube-controller/2.log" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.381807 4896 generic.go:334] "Generic (PLEG): container finished" podID="449d40af-fe23-4350-b66f-de4ce6614177" containerID="670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407" exitCode=1 Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.381848 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" event={"ID":"449d40af-fe23-4350-b66f-de4ce6614177","Type":"ContainerDied","Data":"670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407"} Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.381925 4896 scope.go:117] "RemoveContainer" containerID="3705371cdc4101eed891047c7e68b95155ea083842e0343ab98763e25ea00f73" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.382749 4896 scope.go:117] "RemoveContainer" containerID="670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407" Feb 18 14:56:24 crc kubenswrapper[4896]: E0218 14:56:24.382936 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-w862c_openshift-ovn-kubernetes(449d40af-fe23-4350-b66f-de4ce6614177)\"" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" podUID="449d40af-fe23-4350-b66f-de4ce6614177" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.402103 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.406083 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.406100 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.406107 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.406119 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.406127 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:24Z","lastTransitionTime":"2026-02-18T14:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.417906 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.433562 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6e1900020612d617118e0cca524ffaafe1fac28997fcf1f73a75fe743b4f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.445545 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6wprd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8896bb9-c6fe-4111-a49a-9dadf23e6855\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80d2101411bb762554b416ff237ce9ce2455e43c3c5379118322264ffd9dcd94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5jc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6wprd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.463527 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.477635 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e1cf96a28a07565ec2d56a4ffc32c10e95356b7ebf2399b6b3ba833274a68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.488863 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"219e676d-1cbc-4503-bef3-7b3c5af4e0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd5e328de9a4e016452b309a56258165a94fdd43c0e26f7444a4a765f3cd11c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec861caaac932e4489277e16914fa252647b155492fec4edeb789ea62e11eac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d26b99304cc56b1e016ba622e79f535b6dfd04c197b698dae1720a1ee44afab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc64cbe4646c598d98e96a7ff130d52867c29c552541b1c7adea6530a582e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bdc64cbe4646c598d98e96a7ff130d52867c29c552541b1c7adea6530a582e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.501221 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.508387 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.508435 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.508456 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.508478 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.508495 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:24Z","lastTransitionTime":"2026-02-18T14:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.513138 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.525229 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-98x74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"229213f7-7a25-4982-a449-0122912ba05c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-98x74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.540050 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.551420 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.563369 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.574751 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.582839 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.592301 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32bd695f-cf0b-425a-b764-87b3f678d4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://390187c9005f214d18d43cd6b4f987ccc7d2197f8629997ca4c5cd3ffb49c777\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7021f783fefb51e5fb6b701e58eb9279a0513b9ff37cba435d69e1b7bb1d1a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8j9gb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.604400 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54f94d214d96548b9d5a424bc162f56ffecc188391e16ee0c3b4009d55ca0472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:56:11Z\\\",\\\"message\\\":\\\"2026-02-18T14:55:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_535ee55e-8811-41f2-8919-a177df470c76\\\\n2026-02-18T14:55:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_535ee55e-8811-41f2-8919-a177df470c76 to /host/opt/cni/bin/\\\\n2026-02-18T14:55:26Z [verbose] multus-daemon started\\\\n2026-02-18T14:55:26Z [verbose] Readiness Indicator file check\\\\n2026-02-18T14:56:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.610056 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.610126 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.610141 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.610156 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.610167 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:24Z","lastTransitionTime":"2026-02-18T14:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.623854 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3705371cdc4101eed891047c7e68b95155ea083842e0343ab98763e25ea00f73\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:55:53Z\\\",\\\"message\\\":\\\"ew object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI0218 14:55:53.908904 6577 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nI0218 14:55:53.908922 6577 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0218 14:55:53.908938 6577 base_network_controller_pods.go:477] [default/openshift-network-console/networking-console-plugin-85b44fc459-gdk6g] creating logical port openshift-network-console_networking-console-plugin-85b44fc459-gdk6g for pod on switch crc\\\\nI0218 14:55:53.908947 6577 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-oauth-apiserver/api\\\\\\\"}\\\\nI0218 14:55:53.908958 6577 services_controller.go:360] Finished syncing service api on namespace openshift-oauth-apiserver for network=default : 2.740953ms\\\\nI0218 14:55:53.908969 6577 services_controller.go:356] Processing sync for service openshift-network-diagnostics/network-check-target for network=default\\\\nF0218 14:55:53.908970 6577 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:53Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:56:23Z\\\",\\\"message\\\":\\\"to complete in iterateRetryResources\\\\nI0218 14:56:23.858541 7025 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nF0218 14:56:23.858547 7025 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:23Z is after 2025-08-24T17:21:41Z]\\\\nI0218 14:56:23.858448 7025 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/olm-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster o\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:56:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:24Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.712824 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.712876 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.712886 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.712900 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.712911 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:24Z","lastTransitionTime":"2026-02-18T14:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.814984 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.815032 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.815050 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.815073 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.815090 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:24Z","lastTransitionTime":"2026-02-18T14:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.912774 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-21 16:00:30.937171368 +0000 UTC Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.917267 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.917309 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.917322 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.917340 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.917353 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:24Z","lastTransitionTime":"2026-02-18T14:56:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.927668 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.927711 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:56:24 crc kubenswrapper[4896]: I0218 14:56:24.927723 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:56:24 crc kubenswrapper[4896]: E0218 14:56:24.927799 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:56:24 crc kubenswrapper[4896]: E0218 14:56:24.927899 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:56:24 crc kubenswrapper[4896]: E0218 14:56:24.927995 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.019767 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.019799 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.019809 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.019822 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.019829 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:25Z","lastTransitionTime":"2026-02-18T14:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.123111 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.123154 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.123164 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.123178 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.123186 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:25Z","lastTransitionTime":"2026-02-18T14:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.225117 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.225150 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.225158 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.225172 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.225181 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:25Z","lastTransitionTime":"2026-02-18T14:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.327087 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.327142 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.327151 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.327165 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.327173 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:25Z","lastTransitionTime":"2026-02-18T14:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.386358 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w862c_449d40af-fe23-4350-b66f-de4ce6614177/ovnkube-controller/3.log" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.390066 4896 scope.go:117] "RemoveContainer" containerID="670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407" Feb 18 14:56:25 crc kubenswrapper[4896]: E0218 14:56:25.390299 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-w862c_openshift-ovn-kubernetes(449d40af-fe23-4350-b66f-de4ce6614177)\"" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" podUID="449d40af-fe23-4350-b66f-de4ce6614177" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.404908 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"86d205a2-f5b5-4472-89bc-b640a7a504bf\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da18359b524389bcd4f88dde6e9f7c2b05bb5fce3845df3a9c98fe24ab816322\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d8799131b1cd8e2f5d7517ecc435c43594533844a1b4c920c0a994ff2dc0447f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://09cc2a7a38615b7afff3a26d6494bbca1b473a1955eccaa6cbec1c216b60c02f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.417307 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://144e85ba83c79a72b5f3515ceac99e8a62b3cfa1b4c8b6201f293ba93a20ef08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.430134 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.430249 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.430263 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.430282 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.430298 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:25Z","lastTransitionTime":"2026-02-18T14:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.435120 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://269879b26e56ecc8f201af726b200f55dff096f6bec4852a6879eda29509d887\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b2401dbfd8f6b3efa69a09654616ed5d9516ae6a24de9105b2cc2259dd962bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.450700 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.461652 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bxlt7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3c76e871-21e3-4654-9fbb-9770bf54ee0b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f02b801556e7ad67a5c075c7d08a5066efd7cde07dcb088ffbd09becc8c047f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h2bpm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bxlt7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.473346 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32bd695f-cf0b-425a-b764-87b3f678d4ea\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://390187c9005f214d18d43cd6b4f987ccc7d2197f8629997ca4c5cd3ffb49c777\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f7021f783fefb51e5fb6b701e58eb9279a0513b9ff37cba435d69e1b7bb1d1a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jk59\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8j9gb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.485020 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-98x74" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"229213f7-7a25-4982-a449-0122912ba05c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5rmp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-98x74\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.498974 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-fmg2f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c5ba4981-ee97-4d71-b48f-b5d12bd5911c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:56:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://54f94d214d96548b9d5a424bc162f56ffecc188391e16ee0c3b4009d55ca0472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:56:11Z\\\",\\\"message\\\":\\\"2026-02-18T14:55:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_535ee55e-8811-41f2-8919-a177df470c76\\\\n2026-02-18T14:55:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_535ee55e-8811-41f2-8919-a177df470c76 to /host/opt/cni/bin/\\\\n2026-02-18T14:55:26Z [verbose] multus-daemon started\\\\n2026-02-18T14:55:26Z [verbose] Readiness Indicator file check\\\\n2026-02-18T14:56:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:56:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5rmg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-fmg2f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.517803 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"449d40af-fe23-4350-b66f-de4ce6614177\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-18T14:56:23Z\\\",\\\"message\\\":\\\"to complete in iterateRetryResources\\\\nI0218 14:56:23.858541 7025 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nF0218 14:56:23.858547 7025 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:23Z is after 2025-08-24T17:21:41Z]\\\\nI0218 14:56:23.858448 7025 model_client.go:382] Update operations generated as: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-operator-lifecycle-manager/olm-operator-metrics]} name:Service_openshift-operator-lifecycle-manager/olm-operator-metrics_TCP_cluster o\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:56:23Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-w862c_openshift-ovn-kubernetes(449d40af-fe23-4350-b66f-de4ce6614177)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9zjjt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-w862c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.528482 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.532665 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.532693 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.532704 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.532719 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.532728 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:25Z","lastTransitionTime":"2026-02-18T14:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.539783 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.553282 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-scgl4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2efa13d-b712-4f54-967a-d948be1fce0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6e1900020612d617118e0cca524ffaafe1fac28997fcf1f73a75fe743b4f55\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://68b278ba8565c6eea8a7353c83ef5ce34c4c4d0b6e237bb6635260ccf04cca13\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c17d78ef47682764c5b80010ac1d6e3846dd1bb7f2aa23510b19d266028cf02e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ccc55fd51612c13e519cd42b118f0293f3bc1dd2e9b845f8af73db25dad1a581\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6036a062dfaccce99f0e589568f3f2d29d9e7dcaa4a6012452d0851a085b85e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b71e8440aa6032a3588f2d5583d3f0f19c185edfc7955e818a84a77e1df14e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bb043c8d15290bf4d368b83f3cce8cbbfdd19126cffb26ed020e41bed9c0dd3a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rn6c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-scgl4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.562453 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6wprd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e8896bb9-c6fe-4111-a49a-9dadf23e6855\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80d2101411bb762554b416ff237ce9ce2455e43c3c5379118322264ffd9dcd94\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5jc4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:28Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6wprd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.578079 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de866e5e-85f2-4abb-902a-a1db851f822d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f13bb80150829378d10ebf7cca34fd54f77ad85d7cc943bef1fdd871d7cd7283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b590716db6cc899042843b8ffb16ba518d82fcd9b3b396d35848035f03419e23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d88dbb722ec0f456b3e018a06be8a08438d5b89c9ac55f28cf9119c3ea968b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69713a063ac63f0932f8dc051bda908c96c83fcc392998885674de10f935183\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5468d8418f72b88342ac8912589df6a99bedac49338112b16389cdd682895cb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c4b6f4a8326a672a3f9c3af91e47bd8735d1c280823655fd26b51a52955e0a1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06fa67ee3c1b2a60266ad2b78b5874669b7d56ef138b59ebec835d1331ab69b0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b838d33f7f561f6c410fe567a048aaeaf30c73ced63948725ac0341338c5eb65\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.592212 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e1e5f7db-b3a9-44d9-8ab1-1950995db98a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://646e1cf96a28a07565ec2d56a4ffc32c10e95356b7ebf2399b6b3ba833274a68\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-18T14:55:23Z\\\",\\\"message\\\":\\\"le observer\\\\nW0218 14:55:22.664045 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0218 14:55:22.664168 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0218 14:55:22.664873 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1653022756/tls.crt::/tmp/serving-cert-1653022756/tls.key\\\\\\\"\\\\nI0218 14:55:22.984817 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0218 14:55:22.990545 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0218 14:55:22.994677 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0218 14:55:22.994822 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0218 14:55:22.994871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0218 14:55:22.999848 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0218 14:55:22.999899 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999924 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0218 14:55:22.999946 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0218 14:55:22.999970 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0218 14:55:22.999992 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0218 14:55:23.000012 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0218 14:55:23.000048 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0218 14:55:23.003304 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:17Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.603709 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"219e676d-1cbc-4503-bef3-7b3c5af4e0ab\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd5e328de9a4e016452b309a56258165a94fdd43c0e26f7444a4a765f3cd11c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec861caaac932e4489277e16914fa252647b155492fec4edeb789ea62e11eac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d26b99304cc56b1e016ba622e79f535b6dfd04c197b698dae1720a1ee44afab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3bdc64cbe4646c598d98e96a7ff130d52867c29c552541b1c7adea6530a582e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bdc64cbe4646c598d98e96a7ff130d52867c29c552541b1c7adea6530a582e9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-18T14:55:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-18T14:55:03Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:02Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.615995 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2c19173bc8771fa7fb23ecdf1a51b2cfe496d05ad56c8359c830a346f0c4a73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.627492 4896 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8503734a-8111-4fab-beda-1f0f3b59615b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-18T14:55:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fdd2be74861dc67af37b070967a18455607e83421d64d8fc9e3424b41767811b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-18T14:55:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-n5k2b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-18T14:55:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-24rq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-18T14:56:25Z is after 2025-08-24T17:21:41Z" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.635292 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.635331 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.635347 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.635364 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.635376 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:25Z","lastTransitionTime":"2026-02-18T14:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.737524 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.737560 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.737569 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.737580 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.737588 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:25Z","lastTransitionTime":"2026-02-18T14:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.839179 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.839214 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.839226 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.839262 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.839271 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:25Z","lastTransitionTime":"2026-02-18T14:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.913777 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 02:31:10.442052376 +0000 UTC Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.927151 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:56:25 crc kubenswrapper[4896]: E0218 14:56:25.927316 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.941423 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.941463 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.941475 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.941492 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:25 crc kubenswrapper[4896]: I0218 14:56:25.941503 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:25Z","lastTransitionTime":"2026-02-18T14:56:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.044095 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.044121 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.044129 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.044153 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.044163 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:26Z","lastTransitionTime":"2026-02-18T14:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.146008 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.146039 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.146048 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.146060 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.146068 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:26Z","lastTransitionTime":"2026-02-18T14:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.247714 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.247745 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.247754 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.247769 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.247778 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:26Z","lastTransitionTime":"2026-02-18T14:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.350246 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.350283 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.350293 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.350307 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.350318 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:26Z","lastTransitionTime":"2026-02-18T14:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.453753 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.453801 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.453814 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.453830 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.453844 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:26Z","lastTransitionTime":"2026-02-18T14:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.556091 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.556132 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.556141 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.556154 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.556162 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:26Z","lastTransitionTime":"2026-02-18T14:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.658811 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.658862 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.658878 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.658899 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.658915 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:26Z","lastTransitionTime":"2026-02-18T14:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.761217 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.761266 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.761276 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.761288 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.761297 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:26Z","lastTransitionTime":"2026-02-18T14:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.794095 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:56:26 crc kubenswrapper[4896]: E0218 14:56:26.794221 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:30.794202995 +0000 UTC m=+149.723537208 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.794285 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.794319 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.794337 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.794358 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:56:26 crc kubenswrapper[4896]: E0218 14:56:26.794452 4896 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 14:56:26 crc kubenswrapper[4896]: E0218 14:56:26.794453 4896 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 14:56:26 crc kubenswrapper[4896]: E0218 14:56:26.794529 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 14:57:30.794508843 +0000 UTC m=+149.723843096 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 18 14:56:26 crc kubenswrapper[4896]: E0218 14:56:26.794528 4896 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 14:56:26 crc kubenswrapper[4896]: E0218 14:56:26.794464 4896 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 14:56:26 crc kubenswrapper[4896]: E0218 14:56:26.794584 4896 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 14:56:26 crc kubenswrapper[4896]: E0218 14:56:26.794525 4896 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 18 14:56:26 crc kubenswrapper[4896]: E0218 14:56:26.794666 4896 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 18 14:56:26 crc kubenswrapper[4896]: E0218 14:56:26.794692 4896 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 14:56:26 crc kubenswrapper[4896]: E0218 14:56:26.794594 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-18 14:57:30.794577215 +0000 UTC m=+149.723911428 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 18 14:56:26 crc kubenswrapper[4896]: E0218 14:56:26.794717 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-18 14:57:30.794707458 +0000 UTC m=+149.724041671 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 14:56:26 crc kubenswrapper[4896]: E0218 14:56:26.794774 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-18 14:57:30.794749189 +0000 UTC m=+149.724083412 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.863552 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.863586 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.863594 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.863607 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.863616 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:26Z","lastTransitionTime":"2026-02-18T14:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.914926 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 02:46:05.099860048 +0000 UTC Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.927405 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.927425 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.927467 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:56:26 crc kubenswrapper[4896]: E0218 14:56:26.927601 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:56:26 crc kubenswrapper[4896]: E0218 14:56:26.927681 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:56:26 crc kubenswrapper[4896]: E0218 14:56:26.927722 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.970925 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.970958 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.970966 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.970978 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:26 crc kubenswrapper[4896]: I0218 14:56:26.970986 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:26Z","lastTransitionTime":"2026-02-18T14:56:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.073117 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.073187 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.073205 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.073231 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.073290 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:27Z","lastTransitionTime":"2026-02-18T14:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.175627 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.175688 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.175703 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.175724 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.175737 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:27Z","lastTransitionTime":"2026-02-18T14:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.278044 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.278084 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.278131 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.278149 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.278160 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:27Z","lastTransitionTime":"2026-02-18T14:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.380222 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.380271 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.380280 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.380294 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.380305 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:27Z","lastTransitionTime":"2026-02-18T14:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.482701 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.482752 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.482763 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.482783 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.482797 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:27Z","lastTransitionTime":"2026-02-18T14:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.585388 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.585427 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.585440 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.585459 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.585471 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:27Z","lastTransitionTime":"2026-02-18T14:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.687983 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.688018 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.688027 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.688040 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.688048 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:27Z","lastTransitionTime":"2026-02-18T14:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.791120 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.791166 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.791179 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.791198 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.791211 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:27Z","lastTransitionTime":"2026-02-18T14:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.894027 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.894077 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.894094 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.894112 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.894126 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:27Z","lastTransitionTime":"2026-02-18T14:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.915968 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 16:31:53.349904248 +0000 UTC Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.927350 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:56:27 crc kubenswrapper[4896]: E0218 14:56:27.927462 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.996926 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.996964 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.996979 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.996998 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:27 crc kubenswrapper[4896]: I0218 14:56:27.997012 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:27Z","lastTransitionTime":"2026-02-18T14:56:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.099053 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.099082 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.099091 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.099103 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.099112 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:28Z","lastTransitionTime":"2026-02-18T14:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.200807 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.200841 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.200850 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.200862 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.200870 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:28Z","lastTransitionTime":"2026-02-18T14:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.303799 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.303828 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.303836 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.303850 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.303858 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:28Z","lastTransitionTime":"2026-02-18T14:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.406479 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.406727 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.406739 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.406757 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.406767 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:28Z","lastTransitionTime":"2026-02-18T14:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.509052 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.509090 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.509098 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.509112 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.509124 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:28Z","lastTransitionTime":"2026-02-18T14:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.612511 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.612548 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.612559 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.612575 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.612586 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:28Z","lastTransitionTime":"2026-02-18T14:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.714766 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.714800 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.714809 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.714822 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.714830 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:28Z","lastTransitionTime":"2026-02-18T14:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.817560 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.817602 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.817615 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.817632 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.817644 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:28Z","lastTransitionTime":"2026-02-18T14:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.916490 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 20:37:26.113039282 +0000 UTC Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.920203 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.920366 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.920446 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.920525 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.920588 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:28Z","lastTransitionTime":"2026-02-18T14:56:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.927343 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.927423 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:56:28 crc kubenswrapper[4896]: I0218 14:56:28.927551 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:56:28 crc kubenswrapper[4896]: E0218 14:56:28.927688 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:56:28 crc kubenswrapper[4896]: E0218 14:56:28.927593 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:56:28 crc kubenswrapper[4896]: E0218 14:56:28.927440 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.022403 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.022694 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.022784 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.022863 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.022942 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:29Z","lastTransitionTime":"2026-02-18T14:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.125429 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.125456 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.125467 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.125483 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.125495 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:29Z","lastTransitionTime":"2026-02-18T14:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.228190 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.228221 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.228245 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.228259 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.228267 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:29Z","lastTransitionTime":"2026-02-18T14:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.330565 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.330602 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.330610 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.330626 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.330635 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:29Z","lastTransitionTime":"2026-02-18T14:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.432855 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.432893 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.432906 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.432922 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.432934 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:29Z","lastTransitionTime":"2026-02-18T14:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.535431 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.535516 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.535528 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.535544 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.535561 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:29Z","lastTransitionTime":"2026-02-18T14:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.638676 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.638938 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.639024 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.639153 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.639303 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:29Z","lastTransitionTime":"2026-02-18T14:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.742017 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.742043 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.742051 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.742064 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.742072 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:29Z","lastTransitionTime":"2026-02-18T14:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.844604 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.844635 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.844645 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.844659 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.844668 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:29Z","lastTransitionTime":"2026-02-18T14:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.917637 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-27 10:16:32.210683871 +0000 UTC Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.927248 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:56:29 crc kubenswrapper[4896]: E0218 14:56:29.927505 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.947343 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.947394 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.947406 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.947427 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:29 crc kubenswrapper[4896]: I0218 14:56:29.947439 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:29Z","lastTransitionTime":"2026-02-18T14:56:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.050002 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.050294 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.050396 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.050486 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.050579 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:30Z","lastTransitionTime":"2026-02-18T14:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.152900 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.153187 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.153330 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.153428 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.153523 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:30Z","lastTransitionTime":"2026-02-18T14:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.256031 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.256310 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.256395 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.256617 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.256764 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:30Z","lastTransitionTime":"2026-02-18T14:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.359578 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.359608 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.359617 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.359629 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.359637 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:30Z","lastTransitionTime":"2026-02-18T14:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.462416 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.462469 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.462482 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.462498 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.462509 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:30Z","lastTransitionTime":"2026-02-18T14:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.564705 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.564768 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.564780 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.564796 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.564808 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:30Z","lastTransitionTime":"2026-02-18T14:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.613384 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.613430 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.613441 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.613457 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.613468 4896 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-18T14:56:30Z","lastTransitionTime":"2026-02-18T14:56:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.659165 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-9pmwz"] Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.659500 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9pmwz" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.661441 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.661597 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.661766 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.662850 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.691793 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=68.69177158 podStartE2EDuration="1m8.69177158s" podCreationTimestamp="2026-02-18 14:55:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:56:30.691750689 +0000 UTC m=+89.621084922" watchObservedRunningTime="2026-02-18 14:56:30.69177158 +0000 UTC m=+89.621105803" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.707459 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=67.707438475 podStartE2EDuration="1m7.707438475s" podCreationTimestamp="2026-02-18 14:55:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:56:30.707295401 +0000 UTC m=+89.636629624" watchObservedRunningTime="2026-02-18 14:56:30.707438475 +0000 UTC m=+89.636772688" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.735281 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/858ec096-9fd1-42df-bb8e-1d91565adcab-service-ca\") pod \"cluster-version-operator-5c965bbfc6-9pmwz\" (UID: \"858ec096-9fd1-42df-bb8e-1d91565adcab\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9pmwz" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.735698 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/858ec096-9fd1-42df-bb8e-1d91565adcab-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-9pmwz\" (UID: \"858ec096-9fd1-42df-bb8e-1d91565adcab\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9pmwz" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.735924 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/858ec096-9fd1-42df-bb8e-1d91565adcab-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-9pmwz\" (UID: \"858ec096-9fd1-42df-bb8e-1d91565adcab\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9pmwz" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.736142 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/858ec096-9fd1-42df-bb8e-1d91565adcab-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-9pmwz\" (UID: \"858ec096-9fd1-42df-bb8e-1d91565adcab\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9pmwz" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.736476 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/858ec096-9fd1-42df-bb8e-1d91565adcab-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-9pmwz\" (UID: \"858ec096-9fd1-42df-bb8e-1d91565adcab\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9pmwz" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.763351 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=33.763332327 podStartE2EDuration="33.763332327s" podCreationTimestamp="2026-02-18 14:55:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:56:30.7347642 +0000 UTC m=+89.664098423" watchObservedRunningTime="2026-02-18 14:56:30.763332327 +0000 UTC m=+89.692666540" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.790018 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=68.790001884 podStartE2EDuration="1m8.790001884s" podCreationTimestamp="2026-02-18 14:55:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:56:30.789519901 +0000 UTC m=+89.718854124" watchObservedRunningTime="2026-02-18 14:56:30.790001884 +0000 UTC m=+89.719336097" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.790957 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podStartSLOduration=66.790947909 podStartE2EDuration="1m6.790947909s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:56:30.777259736 +0000 UTC m=+89.706593969" watchObservedRunningTime="2026-02-18 14:56:30.790947909 +0000 UTC m=+89.720282142" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.838037 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/858ec096-9fd1-42df-bb8e-1d91565adcab-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-9pmwz\" (UID: \"858ec096-9fd1-42df-bb8e-1d91565adcab\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9pmwz" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.838078 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/858ec096-9fd1-42df-bb8e-1d91565adcab-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-9pmwz\" (UID: \"858ec096-9fd1-42df-bb8e-1d91565adcab\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9pmwz" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.838109 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/858ec096-9fd1-42df-bb8e-1d91565adcab-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-9pmwz\" (UID: \"858ec096-9fd1-42df-bb8e-1d91565adcab\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9pmwz" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.838149 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/858ec096-9fd1-42df-bb8e-1d91565adcab-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-9pmwz\" (UID: \"858ec096-9fd1-42df-bb8e-1d91565adcab\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9pmwz" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.838165 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/858ec096-9fd1-42df-bb8e-1d91565adcab-service-ca\") pod \"cluster-version-operator-5c965bbfc6-9pmwz\" (UID: \"858ec096-9fd1-42df-bb8e-1d91565adcab\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9pmwz" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.838176 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/858ec096-9fd1-42df-bb8e-1d91565adcab-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-9pmwz\" (UID: \"858ec096-9fd1-42df-bb8e-1d91565adcab\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9pmwz" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.838288 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/858ec096-9fd1-42df-bb8e-1d91565adcab-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-9pmwz\" (UID: \"858ec096-9fd1-42df-bb8e-1d91565adcab\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9pmwz" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.838976 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/858ec096-9fd1-42df-bb8e-1d91565adcab-service-ca\") pod \"cluster-version-operator-5c965bbfc6-9pmwz\" (UID: \"858ec096-9fd1-42df-bb8e-1d91565adcab\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9pmwz" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.839654 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-bxlt7" podStartSLOduration=66.83964431 podStartE2EDuration="1m6.83964431s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:56:30.839284481 +0000 UTC m=+89.768618694" watchObservedRunningTime="2026-02-18 14:56:30.83964431 +0000 UTC m=+89.768978523" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.842973 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/858ec096-9fd1-42df-bb8e-1d91565adcab-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-9pmwz\" (UID: \"858ec096-9fd1-42df-bb8e-1d91565adcab\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9pmwz" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.851265 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8j9gb" podStartSLOduration=66.851222337 podStartE2EDuration="1m6.851222337s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:56:30.850753175 +0000 UTC m=+89.780087388" watchObservedRunningTime="2026-02-18 14:56:30.851222337 +0000 UTC m=+89.780556550" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.861437 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/858ec096-9fd1-42df-bb8e-1d91565adcab-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-9pmwz\" (UID: \"858ec096-9fd1-42df-bb8e-1d91565adcab\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9pmwz" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.876692 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-fmg2f" podStartSLOduration=66.876675372 podStartE2EDuration="1m6.876675372s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:56:30.87656941 +0000 UTC m=+89.805903623" watchObservedRunningTime="2026-02-18 14:56:30.876675372 +0000 UTC m=+89.806009585" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.918136 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 23:06:54.190804183 +0000 UTC Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.918201 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.924781 4896 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.926897 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.926948 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:56:30 crc kubenswrapper[4896]: E0218 14:56:30.927012 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.926954 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:56:30 crc kubenswrapper[4896]: E0218 14:56:30.927089 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:56:30 crc kubenswrapper[4896]: E0218 14:56:30.927267 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.949614 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-scgl4" podStartSLOduration=66.949592026 podStartE2EDuration="1m6.949592026s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:56:30.948767634 +0000 UTC m=+89.878101847" watchObservedRunningTime="2026-02-18 14:56:30.949592026 +0000 UTC m=+89.878926239" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.961708 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-6wprd" podStartSLOduration=66.961692377 podStartE2EDuration="1m6.961692377s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:56:30.961521322 +0000 UTC m=+89.890855545" watchObservedRunningTime="2026-02-18 14:56:30.961692377 +0000 UTC m=+89.891026590" Feb 18 14:56:30 crc kubenswrapper[4896]: I0218 14:56:30.974032 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9pmwz" Feb 18 14:56:31 crc kubenswrapper[4896]: I0218 14:56:31.405715 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9pmwz" event={"ID":"858ec096-9fd1-42df-bb8e-1d91565adcab","Type":"ContainerStarted","Data":"a01d7b3d7790722dcf93f1d25c467ee8bbab1037ba4d1b45c32ea0270cd28495"} Feb 18 14:56:31 crc kubenswrapper[4896]: I0218 14:56:31.405758 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9pmwz" event={"ID":"858ec096-9fd1-42df-bb8e-1d91565adcab","Type":"ContainerStarted","Data":"c939efdf7abb4b37074935a8bcbe6b36896870e649f198cd4902a309dc37fecd"} Feb 18 14:56:31 crc kubenswrapper[4896]: I0218 14:56:31.422300 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9pmwz" podStartSLOduration=67.422207457 podStartE2EDuration="1m7.422207457s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:56:31.418787937 +0000 UTC m=+90.348122180" watchObservedRunningTime="2026-02-18 14:56:31.422207457 +0000 UTC m=+90.351541670" Feb 18 14:56:31 crc kubenswrapper[4896]: I0218 14:56:31.927584 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:56:31 crc kubenswrapper[4896]: E0218 14:56:31.927948 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:56:31 crc kubenswrapper[4896]: I0218 14:56:31.938108 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Feb 18 14:56:32 crc kubenswrapper[4896]: I0218 14:56:32.927368 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:56:32 crc kubenswrapper[4896]: I0218 14:56:32.927429 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:56:32 crc kubenswrapper[4896]: E0218 14:56:32.927626 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:56:32 crc kubenswrapper[4896]: I0218 14:56:32.927756 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:56:32 crc kubenswrapper[4896]: E0218 14:56:32.927829 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:56:32 crc kubenswrapper[4896]: E0218 14:56:32.928135 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:56:33 crc kubenswrapper[4896]: I0218 14:56:33.937124 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:56:33 crc kubenswrapper[4896]: E0218 14:56:33.937314 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:56:34 crc kubenswrapper[4896]: I0218 14:56:34.927190 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:56:34 crc kubenswrapper[4896]: E0218 14:56:34.927566 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:56:34 crc kubenswrapper[4896]: I0218 14:56:34.927190 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:56:34 crc kubenswrapper[4896]: I0218 14:56:34.927190 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:56:34 crc kubenswrapper[4896]: E0218 14:56:34.927626 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:56:34 crc kubenswrapper[4896]: E0218 14:56:34.927696 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:56:35 crc kubenswrapper[4896]: I0218 14:56:35.927882 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:56:35 crc kubenswrapper[4896]: E0218 14:56:35.928012 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:56:36 crc kubenswrapper[4896]: I0218 14:56:36.927723 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:56:36 crc kubenswrapper[4896]: I0218 14:56:36.927723 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:56:36 crc kubenswrapper[4896]: I0218 14:56:36.927740 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:56:36 crc kubenswrapper[4896]: E0218 14:56:36.928368 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:56:36 crc kubenswrapper[4896]: E0218 14:56:36.928489 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:56:36 crc kubenswrapper[4896]: E0218 14:56:36.928607 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:56:37 crc kubenswrapper[4896]: I0218 14:56:37.928535 4896 scope.go:117] "RemoveContainer" containerID="670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407" Feb 18 14:56:37 crc kubenswrapper[4896]: E0218 14:56:37.928677 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-w862c_openshift-ovn-kubernetes(449d40af-fe23-4350-b66f-de4ce6614177)\"" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" podUID="449d40af-fe23-4350-b66f-de4ce6614177" Feb 18 14:56:37 crc kubenswrapper[4896]: I0218 14:56:37.928923 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:56:37 crc kubenswrapper[4896]: E0218 14:56:37.928996 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:56:38 crc kubenswrapper[4896]: I0218 14:56:38.927770 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:56:38 crc kubenswrapper[4896]: I0218 14:56:38.927819 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:56:38 crc kubenswrapper[4896]: E0218 14:56:38.927909 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:56:38 crc kubenswrapper[4896]: E0218 14:56:38.928053 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:56:38 crc kubenswrapper[4896]: I0218 14:56:38.929449 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:56:38 crc kubenswrapper[4896]: E0218 14:56:38.929559 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:56:39 crc kubenswrapper[4896]: I0218 14:56:39.927801 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:56:39 crc kubenswrapper[4896]: E0218 14:56:39.928002 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:56:40 crc kubenswrapper[4896]: I0218 14:56:40.927364 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:56:40 crc kubenswrapper[4896]: I0218 14:56:40.927380 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:56:40 crc kubenswrapper[4896]: I0218 14:56:40.927439 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:56:40 crc kubenswrapper[4896]: E0218 14:56:40.927472 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:56:40 crc kubenswrapper[4896]: E0218 14:56:40.927555 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:56:40 crc kubenswrapper[4896]: E0218 14:56:40.927671 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:56:41 crc kubenswrapper[4896]: I0218 14:56:41.927917 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:56:41 crc kubenswrapper[4896]: E0218 14:56:41.928999 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:56:41 crc kubenswrapper[4896]: I0218 14:56:41.939873 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=10.939858763 podStartE2EDuration="10.939858763s" podCreationTimestamp="2026-02-18 14:56:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:56:41.938765164 +0000 UTC m=+100.868099377" watchObservedRunningTime="2026-02-18 14:56:41.939858763 +0000 UTC m=+100.869192976" Feb 18 14:56:42 crc kubenswrapper[4896]: I0218 14:56:42.351252 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/229213f7-7a25-4982-a449-0122912ba05c-metrics-certs\") pod \"network-metrics-daemon-98x74\" (UID: \"229213f7-7a25-4982-a449-0122912ba05c\") " pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:56:42 crc kubenswrapper[4896]: E0218 14:56:42.351390 4896 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 14:56:42 crc kubenswrapper[4896]: E0218 14:56:42.351468 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/229213f7-7a25-4982-a449-0122912ba05c-metrics-certs podName:229213f7-7a25-4982-a449-0122912ba05c nodeName:}" failed. No retries permitted until 2026-02-18 14:57:46.351451446 +0000 UTC m=+165.280785659 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/229213f7-7a25-4982-a449-0122912ba05c-metrics-certs") pod "network-metrics-daemon-98x74" (UID: "229213f7-7a25-4982-a449-0122912ba05c") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 18 14:56:42 crc kubenswrapper[4896]: I0218 14:56:42.927447 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:56:42 crc kubenswrapper[4896]: I0218 14:56:42.927515 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:56:42 crc kubenswrapper[4896]: E0218 14:56:42.927565 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:56:42 crc kubenswrapper[4896]: E0218 14:56:42.927674 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:56:42 crc kubenswrapper[4896]: I0218 14:56:42.927685 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:56:42 crc kubenswrapper[4896]: E0218 14:56:42.927885 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:56:43 crc kubenswrapper[4896]: I0218 14:56:43.927573 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:56:43 crc kubenswrapper[4896]: E0218 14:56:43.927670 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:56:44 crc kubenswrapper[4896]: I0218 14:56:44.927002 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:56:44 crc kubenswrapper[4896]: I0218 14:56:44.927009 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:56:44 crc kubenswrapper[4896]: E0218 14:56:44.927189 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:56:44 crc kubenswrapper[4896]: I0218 14:56:44.927015 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:56:44 crc kubenswrapper[4896]: E0218 14:56:44.927311 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:56:44 crc kubenswrapper[4896]: E0218 14:56:44.927382 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:56:45 crc kubenswrapper[4896]: I0218 14:56:45.928026 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:56:45 crc kubenswrapper[4896]: E0218 14:56:45.928198 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:56:46 crc kubenswrapper[4896]: I0218 14:56:46.927158 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:56:46 crc kubenswrapper[4896]: E0218 14:56:46.927302 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:56:46 crc kubenswrapper[4896]: I0218 14:56:46.927319 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:56:46 crc kubenswrapper[4896]: I0218 14:56:46.927376 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:56:46 crc kubenswrapper[4896]: E0218 14:56:46.927461 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:56:46 crc kubenswrapper[4896]: E0218 14:56:46.927624 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:56:47 crc kubenswrapper[4896]: I0218 14:56:47.927802 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:56:47 crc kubenswrapper[4896]: E0218 14:56:47.927997 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:56:48 crc kubenswrapper[4896]: I0218 14:56:48.927834 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:56:48 crc kubenswrapper[4896]: I0218 14:56:48.927860 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:56:48 crc kubenswrapper[4896]: I0218 14:56:48.927834 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:56:48 crc kubenswrapper[4896]: E0218 14:56:48.927936 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:56:48 crc kubenswrapper[4896]: E0218 14:56:48.928061 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:56:48 crc kubenswrapper[4896]: E0218 14:56:48.928157 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:56:49 crc kubenswrapper[4896]: I0218 14:56:49.927679 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:56:49 crc kubenswrapper[4896]: E0218 14:56:49.927932 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:56:50 crc kubenswrapper[4896]: I0218 14:56:50.927293 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:56:50 crc kubenswrapper[4896]: I0218 14:56:50.927391 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:56:50 crc kubenswrapper[4896]: I0218 14:56:50.927314 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:56:50 crc kubenswrapper[4896]: E0218 14:56:50.927420 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:56:50 crc kubenswrapper[4896]: E0218 14:56:50.927524 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:56:50 crc kubenswrapper[4896]: E0218 14:56:50.927638 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:56:51 crc kubenswrapper[4896]: I0218 14:56:51.928501 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:56:51 crc kubenswrapper[4896]: E0218 14:56:51.928617 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:56:52 crc kubenswrapper[4896]: I0218 14:56:52.927406 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:56:52 crc kubenswrapper[4896]: I0218 14:56:52.927406 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:56:52 crc kubenswrapper[4896]: I0218 14:56:52.927417 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:56:52 crc kubenswrapper[4896]: E0218 14:56:52.928050 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:56:52 crc kubenswrapper[4896]: E0218 14:56:52.928206 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:56:52 crc kubenswrapper[4896]: E0218 14:56:52.928386 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:56:52 crc kubenswrapper[4896]: I0218 14:56:52.928472 4896 scope.go:117] "RemoveContainer" containerID="670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407" Feb 18 14:56:52 crc kubenswrapper[4896]: E0218 14:56:52.928715 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-w862c_openshift-ovn-kubernetes(449d40af-fe23-4350-b66f-de4ce6614177)\"" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" podUID="449d40af-fe23-4350-b66f-de4ce6614177" Feb 18 14:56:53 crc kubenswrapper[4896]: I0218 14:56:53.927433 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:56:53 crc kubenswrapper[4896]: E0218 14:56:53.927595 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:56:54 crc kubenswrapper[4896]: I0218 14:56:54.927190 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:56:54 crc kubenswrapper[4896]: I0218 14:56:54.927274 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:56:54 crc kubenswrapper[4896]: I0218 14:56:54.927190 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:56:54 crc kubenswrapper[4896]: E0218 14:56:54.927350 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:56:54 crc kubenswrapper[4896]: E0218 14:56:54.927407 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:56:54 crc kubenswrapper[4896]: E0218 14:56:54.927469 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:56:55 crc kubenswrapper[4896]: I0218 14:56:55.928473 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:56:55 crc kubenswrapper[4896]: E0218 14:56:55.929257 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:56:56 crc kubenswrapper[4896]: I0218 14:56:56.927965 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:56:56 crc kubenswrapper[4896]: I0218 14:56:56.927976 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:56:56 crc kubenswrapper[4896]: I0218 14:56:56.927990 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:56:56 crc kubenswrapper[4896]: E0218 14:56:56.928289 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:56:56 crc kubenswrapper[4896]: E0218 14:56:56.928385 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:56:56 crc kubenswrapper[4896]: E0218 14:56:56.928439 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:56:57 crc kubenswrapper[4896]: I0218 14:56:57.929405 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:56:57 crc kubenswrapper[4896]: E0218 14:56:57.929743 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:56:58 crc kubenswrapper[4896]: I0218 14:56:58.480181 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fmg2f_c5ba4981-ee97-4d71-b48f-b5d12bd5911c/kube-multus/1.log" Feb 18 14:56:58 crc kubenswrapper[4896]: I0218 14:56:58.480680 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fmg2f_c5ba4981-ee97-4d71-b48f-b5d12bd5911c/kube-multus/0.log" Feb 18 14:56:58 crc kubenswrapper[4896]: I0218 14:56:58.480727 4896 generic.go:334] "Generic (PLEG): container finished" podID="c5ba4981-ee97-4d71-b48f-b5d12bd5911c" containerID="54f94d214d96548b9d5a424bc162f56ffecc188391e16ee0c3b4009d55ca0472" exitCode=1 Feb 18 14:56:58 crc kubenswrapper[4896]: I0218 14:56:58.480788 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fmg2f" event={"ID":"c5ba4981-ee97-4d71-b48f-b5d12bd5911c","Type":"ContainerDied","Data":"54f94d214d96548b9d5a424bc162f56ffecc188391e16ee0c3b4009d55ca0472"} Feb 18 14:56:58 crc kubenswrapper[4896]: I0218 14:56:58.480941 4896 scope.go:117] "RemoveContainer" containerID="434bded82578dfc923377fe4b40671df510405dd9d220c590140244e7fd18b39" Feb 18 14:56:58 crc kubenswrapper[4896]: I0218 14:56:58.481475 4896 scope.go:117] "RemoveContainer" containerID="54f94d214d96548b9d5a424bc162f56ffecc188391e16ee0c3b4009d55ca0472" Feb 18 14:56:58 crc kubenswrapper[4896]: E0218 14:56:58.481681 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-fmg2f_openshift-multus(c5ba4981-ee97-4d71-b48f-b5d12bd5911c)\"" pod="openshift-multus/multus-fmg2f" podUID="c5ba4981-ee97-4d71-b48f-b5d12bd5911c" Feb 18 14:56:58 crc kubenswrapper[4896]: I0218 14:56:58.927577 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:56:58 crc kubenswrapper[4896]: I0218 14:56:58.927651 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:56:58 crc kubenswrapper[4896]: I0218 14:56:58.927711 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:56:58 crc kubenswrapper[4896]: E0218 14:56:58.927774 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:56:58 crc kubenswrapper[4896]: E0218 14:56:58.927871 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:56:58 crc kubenswrapper[4896]: E0218 14:56:58.927962 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:56:59 crc kubenswrapper[4896]: I0218 14:56:59.485002 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fmg2f_c5ba4981-ee97-4d71-b48f-b5d12bd5911c/kube-multus/1.log" Feb 18 14:56:59 crc kubenswrapper[4896]: I0218 14:56:59.927266 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:56:59 crc kubenswrapper[4896]: E0218 14:56:59.927389 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:57:00 crc kubenswrapper[4896]: I0218 14:57:00.927418 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:57:00 crc kubenswrapper[4896]: I0218 14:57:00.927470 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:57:00 crc kubenswrapper[4896]: I0218 14:57:00.927502 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:57:00 crc kubenswrapper[4896]: E0218 14:57:00.927557 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:57:00 crc kubenswrapper[4896]: E0218 14:57:00.927639 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:57:00 crc kubenswrapper[4896]: E0218 14:57:00.927765 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:57:01 crc kubenswrapper[4896]: E0218 14:57:01.888371 4896 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Feb 18 14:57:01 crc kubenswrapper[4896]: I0218 14:57:01.927658 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:57:01 crc kubenswrapper[4896]: E0218 14:57:01.929924 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:57:02 crc kubenswrapper[4896]: E0218 14:57:02.065439 4896 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 18 14:57:02 crc kubenswrapper[4896]: I0218 14:57:02.927056 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:57:02 crc kubenswrapper[4896]: E0218 14:57:02.927624 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:57:02 crc kubenswrapper[4896]: I0218 14:57:02.927181 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:57:02 crc kubenswrapper[4896]: E0218 14:57:02.927900 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:57:02 crc kubenswrapper[4896]: I0218 14:57:02.927077 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:57:02 crc kubenswrapper[4896]: E0218 14:57:02.928016 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:57:03 crc kubenswrapper[4896]: I0218 14:57:03.927336 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:57:03 crc kubenswrapper[4896]: E0218 14:57:03.927503 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:57:04 crc kubenswrapper[4896]: I0218 14:57:04.927103 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:57:04 crc kubenswrapper[4896]: I0218 14:57:04.927177 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:57:04 crc kubenswrapper[4896]: E0218 14:57:04.927264 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:57:04 crc kubenswrapper[4896]: I0218 14:57:04.927294 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:57:04 crc kubenswrapper[4896]: E0218 14:57:04.927307 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:57:04 crc kubenswrapper[4896]: E0218 14:57:04.927614 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:57:04 crc kubenswrapper[4896]: I0218 14:57:04.928013 4896 scope.go:117] "RemoveContainer" containerID="670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407" Feb 18 14:57:05 crc kubenswrapper[4896]: I0218 14:57:05.507206 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w862c_449d40af-fe23-4350-b66f-de4ce6614177/ovnkube-controller/3.log" Feb 18 14:57:05 crc kubenswrapper[4896]: I0218 14:57:05.509047 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" event={"ID":"449d40af-fe23-4350-b66f-de4ce6614177","Type":"ContainerStarted","Data":"4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30"} Feb 18 14:57:05 crc kubenswrapper[4896]: I0218 14:57:05.510168 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:57:05 crc kubenswrapper[4896]: I0218 14:57:05.535736 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" podStartSLOduration=101.535721299 podStartE2EDuration="1m41.535721299s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:05.535601775 +0000 UTC m=+124.464935988" watchObservedRunningTime="2026-02-18 14:57:05.535721299 +0000 UTC m=+124.465055512" Feb 18 14:57:05 crc kubenswrapper[4896]: I0218 14:57:05.715787 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-98x74"] Feb 18 14:57:05 crc kubenswrapper[4896]: I0218 14:57:05.715885 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:57:05 crc kubenswrapper[4896]: E0218 14:57:05.715970 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:57:06 crc kubenswrapper[4896]: I0218 14:57:06.926888 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:57:06 crc kubenswrapper[4896]: I0218 14:57:06.926888 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:57:06 crc kubenswrapper[4896]: E0218 14:57:06.927294 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:57:06 crc kubenswrapper[4896]: E0218 14:57:06.927315 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:57:06 crc kubenswrapper[4896]: I0218 14:57:06.926915 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:57:06 crc kubenswrapper[4896]: E0218 14:57:06.927370 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:57:07 crc kubenswrapper[4896]: E0218 14:57:07.067020 4896 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 18 14:57:07 crc kubenswrapper[4896]: I0218 14:57:07.928092 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:57:07 crc kubenswrapper[4896]: E0218 14:57:07.928303 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:57:08 crc kubenswrapper[4896]: I0218 14:57:08.927956 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:57:08 crc kubenswrapper[4896]: I0218 14:57:08.928013 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:57:08 crc kubenswrapper[4896]: I0218 14:57:08.928118 4896 scope.go:117] "RemoveContainer" containerID="54f94d214d96548b9d5a424bc162f56ffecc188391e16ee0c3b4009d55ca0472" Feb 18 14:57:08 crc kubenswrapper[4896]: I0218 14:57:08.928103 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:57:08 crc kubenswrapper[4896]: E0218 14:57:08.929120 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:57:08 crc kubenswrapper[4896]: E0218 14:57:08.929196 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:57:08 crc kubenswrapper[4896]: E0218 14:57:08.929303 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:57:09 crc kubenswrapper[4896]: I0218 14:57:09.525426 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fmg2f_c5ba4981-ee97-4d71-b48f-b5d12bd5911c/kube-multus/1.log" Feb 18 14:57:09 crc kubenswrapper[4896]: I0218 14:57:09.525537 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fmg2f" event={"ID":"c5ba4981-ee97-4d71-b48f-b5d12bd5911c","Type":"ContainerStarted","Data":"706e92310ef50b2beef6b34b867bc8cb63798490a8311ab1f91116a32f253cd1"} Feb 18 14:57:09 crc kubenswrapper[4896]: I0218 14:57:09.927627 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:57:09 crc kubenswrapper[4896]: E0218 14:57:09.927747 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:57:10 crc kubenswrapper[4896]: I0218 14:57:10.927666 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:57:10 crc kubenswrapper[4896]: I0218 14:57:10.927781 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:57:10 crc kubenswrapper[4896]: E0218 14:57:10.927887 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 18 14:57:10 crc kubenswrapper[4896]: E0218 14:57:10.928090 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 18 14:57:10 crc kubenswrapper[4896]: I0218 14:57:10.928371 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:57:10 crc kubenswrapper[4896]: E0218 14:57:10.928517 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 18 14:57:11 crc kubenswrapper[4896]: I0218 14:57:11.927800 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:57:11 crc kubenswrapper[4896]: E0218 14:57:11.928703 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-98x74" podUID="229213f7-7a25-4982-a449-0122912ba05c" Feb 18 14:57:12 crc kubenswrapper[4896]: I0218 14:57:12.927536 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:57:12 crc kubenswrapper[4896]: I0218 14:57:12.927655 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:57:12 crc kubenswrapper[4896]: I0218 14:57:12.927794 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:57:12 crc kubenswrapper[4896]: I0218 14:57:12.931110 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 18 14:57:12 crc kubenswrapper[4896]: I0218 14:57:12.932530 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 18 14:57:12 crc kubenswrapper[4896]: I0218 14:57:12.933408 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 18 14:57:12 crc kubenswrapper[4896]: I0218 14:57:12.933512 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 18 14:57:13 crc kubenswrapper[4896]: I0218 14:57:13.927280 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:57:13 crc kubenswrapper[4896]: I0218 14:57:13.930437 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 18 14:57:13 crc kubenswrapper[4896]: I0218 14:57:13.931909 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.812930 4896 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.868214 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-22f85"] Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.870441 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-22f85" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.871204 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-v6thv"] Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.878595 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-t2wdv"] Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.879056 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t2wdv" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.879113 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v6thv" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.880041 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr"] Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.880693 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.895826 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.895935 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.896356 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.896375 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.896591 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-6v4cn"] Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.897368 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-6v4cn" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.898602 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ftds"] Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.899089 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ftds" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.900003 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-n25dz"] Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.900433 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bw75d"] Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.901012 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bw75d" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.901461 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-n25dz" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.901733 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-27bbf"] Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.902400 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.902600 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.902808 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.902946 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.904384 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.905473 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.905730 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.906044 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.906222 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.906532 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.906664 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.906785 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.906676 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.906872 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.906918 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.906952 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.906976 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.907008 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.907151 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.907157 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.907179 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-27bbf" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.907269 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.907582 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.907648 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.907735 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.907780 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.909664 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-tgwt9"] Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.910287 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.910518 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-p7zgl"] Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.910618 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.911182 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-p7zgl" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.912601 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.912887 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.913026 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.913173 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.913365 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.913532 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.913690 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.913710 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gqb6w"] Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.913821 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.913890 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.913915 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.916825 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhgdz"] Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.917183 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-mcjhx"] Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.917422 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gqb6w" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.917705 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mcjhx" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.918063 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhgdz" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.923180 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.923220 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.923398 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.923477 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.923584 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.924096 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.924178 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.924281 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.924457 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.926481 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-shvc6"] Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.926961 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.927217 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-4bxkk"] Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.927659 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-4bxkk" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.928462 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.928642 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.928908 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fjmq5"] Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.928948 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.929126 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.929295 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fjmq5" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.929315 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.929442 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.929547 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.929603 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.929553 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.931536 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.931808 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.957679 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.965263 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.966849 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/81190a62-b874-433f-8e50-f6dfe001f6d2-audit-policies\") pod \"apiserver-7bbb656c7d-b8czr\" (UID: \"81190a62-b874-433f-8e50-f6dfe001f6d2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.966883 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2da710c-c8cd-4e5f-8ba3-045858050ade-config\") pod \"machine-approver-56656f9798-t2wdv\" (UID: \"f2da710c-c8cd-4e5f-8ba3-045858050ade\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t2wdv" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.966901 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/81190a62-b874-433f-8e50-f6dfe001f6d2-etcd-client\") pod \"apiserver-7bbb656c7d-b8czr\" (UID: \"81190a62-b874-433f-8e50-f6dfe001f6d2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.966916 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/81190a62-b874-433f-8e50-f6dfe001f6d2-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-b8czr\" (UID: \"81190a62-b874-433f-8e50-f6dfe001f6d2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.966934 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0c6cb779-d660-45f8-8fb2-5829d42ee030-client-ca\") pod \"route-controller-manager-6576b87f9c-v6thv\" (UID: \"0c6cb779-d660-45f8-8fb2-5829d42ee030\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v6thv" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.966947 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81190a62-b874-433f-8e50-f6dfe001f6d2-serving-cert\") pod \"apiserver-7bbb656c7d-b8czr\" (UID: \"81190a62-b874-433f-8e50-f6dfe001f6d2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.966962 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/42ba2ce9-d014-4090-b4b2-21a2ae8cea4c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-22f85\" (UID: \"42ba2ce9-d014-4090-b4b2-21a2ae8cea4c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-22f85" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.966981 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42ba2ce9-d014-4090-b4b2-21a2ae8cea4c-config\") pod \"machine-api-operator-5694c8668f-22f85\" (UID: \"42ba2ce9-d014-4090-b4b2-21a2ae8cea4c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-22f85" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.966995 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f2da710c-c8cd-4e5f-8ba3-045858050ade-machine-approver-tls\") pod \"machine-approver-56656f9798-t2wdv\" (UID: \"f2da710c-c8cd-4e5f-8ba3-045858050ade\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t2wdv" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.967009 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkt2f\" (UniqueName: \"kubernetes.io/projected/42ba2ce9-d014-4090-b4b2-21a2ae8cea4c-kube-api-access-kkt2f\") pod \"machine-api-operator-5694c8668f-22f85\" (UID: \"42ba2ce9-d014-4090-b4b2-21a2ae8cea4c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-22f85" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.967031 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0c6cb779-d660-45f8-8fb2-5829d42ee030-serving-cert\") pod \"route-controller-manager-6576b87f9c-v6thv\" (UID: \"0c6cb779-d660-45f8-8fb2-5829d42ee030\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v6thv" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.967045 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/81190a62-b874-433f-8e50-f6dfe001f6d2-encryption-config\") pod \"apiserver-7bbb656c7d-b8czr\" (UID: \"81190a62-b874-433f-8e50-f6dfe001f6d2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.967058 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/42ba2ce9-d014-4090-b4b2-21a2ae8cea4c-images\") pod \"machine-api-operator-5694c8668f-22f85\" (UID: \"42ba2ce9-d014-4090-b4b2-21a2ae8cea4c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-22f85" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.967081 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c6cb779-d660-45f8-8fb2-5829d42ee030-config\") pod \"route-controller-manager-6576b87f9c-v6thv\" (UID: \"0c6cb779-d660-45f8-8fb2-5829d42ee030\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v6thv" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.967096 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/81190a62-b874-433f-8e50-f6dfe001f6d2-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-b8czr\" (UID: \"81190a62-b874-433f-8e50-f6dfe001f6d2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.967109 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpghg\" (UniqueName: \"kubernetes.io/projected/81190a62-b874-433f-8e50-f6dfe001f6d2-kube-api-access-lpghg\") pod \"apiserver-7bbb656c7d-b8czr\" (UID: \"81190a62-b874-433f-8e50-f6dfe001f6d2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.967124 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/81190a62-b874-433f-8e50-f6dfe001f6d2-audit-dir\") pod \"apiserver-7bbb656c7d-b8czr\" (UID: \"81190a62-b874-433f-8e50-f6dfe001f6d2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.967141 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f2da710c-c8cd-4e5f-8ba3-045858050ade-auth-proxy-config\") pod \"machine-approver-56656f9798-t2wdv\" (UID: \"f2da710c-c8cd-4e5f-8ba3-045858050ade\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t2wdv" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.967167 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhkrn\" (UniqueName: \"kubernetes.io/projected/0c6cb779-d660-45f8-8fb2-5829d42ee030-kube-api-access-hhkrn\") pod \"route-controller-manager-6576b87f9c-v6thv\" (UID: \"0c6cb779-d660-45f8-8fb2-5829d42ee030\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v6thv" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.967192 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4lss\" (UniqueName: \"kubernetes.io/projected/f2da710c-c8cd-4e5f-8ba3-045858050ade-kube-api-access-j4lss\") pod \"machine-approver-56656f9798-t2wdv\" (UID: \"f2da710c-c8cd-4e5f-8ba3-045858050ade\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t2wdv" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.968366 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.968761 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.968906 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.969130 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.969348 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.969404 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-6v4cn"] Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.969425 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-22f85"] Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.969436 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-rmc7c"] Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.969449 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.969706 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.969905 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-z7xdz"] Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.970020 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.970162 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.970292 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-z7xdz" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.970483 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.970490 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-rmc7c" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.970846 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.970968 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.971200 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.971229 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.971324 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.971534 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.971588 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.971645 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.972044 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.972139 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.972412 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.972593 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.972838 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.973457 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-8cq46"] Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.974173 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-8cq46" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.975948 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.976784 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.977039 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.977340 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.977491 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.977607 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.977710 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.977969 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.978023 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-v6thv"] Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.978116 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qrwnx"] Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.978428 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.981537 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.981929 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-6frkl"] Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.982451 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6frkl" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.982530 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.985632 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.986372 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.986508 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-966wr"] Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.987052 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-966wr" Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.987450 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-dlghr"] Feb 18 14:57:21 crc kubenswrapper[4896]: I0218 14:57:21.988226 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dlghr" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.026580 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.027003 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xrhcj"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.027027 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.028189 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xrhcj" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.038965 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ftds"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.039722 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.039891 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.054301 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-pclrg"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.054892 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.054939 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pclrg" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.055227 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.056865 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.058297 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bncs"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.058898 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-fhpqt"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.059371 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-fhpqt" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.059689 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bncs" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.060311 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bjbsw"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.061155 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bjbsw" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.060430 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.062569 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hv6vt"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.063227 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hv6vt" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.064557 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.069433 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42ba2ce9-d014-4090-b4b2-21a2ae8cea4c-config\") pod \"machine-api-operator-5694c8668f-22f85\" (UID: \"42ba2ce9-d014-4090-b4b2-21a2ae8cea4c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-22f85" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.069508 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fd2f0b92-9f4d-4c2c-b83f-7530706cfa09-service-ca-bundle\") pod \"authentication-operator-69f744f599-n25dz\" (UID: \"fd2f0b92-9f4d-4c2c-b83f-7530706cfa09\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-n25dz" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.069540 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f2da710c-c8cd-4e5f-8ba3-045858050ade-machine-approver-tls\") pod \"machine-approver-56656f9798-t2wdv\" (UID: \"f2da710c-c8cd-4e5f-8ba3-045858050ade\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t2wdv" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.069591 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7a51266e-326e-4b5c-a25a-fb61bf1f3757-service-ca-bundle\") pod \"router-default-5444994796-8cq46\" (UID: \"7a51266e-326e-4b5c-a25a-fb61bf1f3757\") " pod="openshift-ingress/router-default-5444994796-8cq46" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.069619 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkt2f\" (UniqueName: \"kubernetes.io/projected/42ba2ce9-d014-4090-b4b2-21a2ae8cea4c-kube-api-access-kkt2f\") pod \"machine-api-operator-5694c8668f-22f85\" (UID: \"42ba2ce9-d014-4090-b4b2-21a2ae8cea4c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-22f85" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.069640 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0c6cb779-d660-45f8-8fb2-5829d42ee030-serving-cert\") pod \"route-controller-manager-6576b87f9c-v6thv\" (UID: \"0c6cb779-d660-45f8-8fb2-5829d42ee030\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v6thv" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.069690 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/81190a62-b874-433f-8e50-f6dfe001f6d2-encryption-config\") pod \"apiserver-7bbb656c7d-b8czr\" (UID: \"81190a62-b874-433f-8e50-f6dfe001f6d2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.069713 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f8cbed1-6086-4fb6-bab5-0563bbcb751c-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-966wr\" (UID: \"7f8cbed1-6086-4fb6-bab5-0563bbcb751c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-966wr" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.069834 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/42ba2ce9-d014-4090-b4b2-21a2ae8cea4c-images\") pod \"machine-api-operator-5694c8668f-22f85\" (UID: \"42ba2ce9-d014-4090-b4b2-21a2ae8cea4c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-22f85" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.069860 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pm7zf\" (UniqueName: \"kubernetes.io/projected/fd2f0b92-9f4d-4c2c-b83f-7530706cfa09-kube-api-access-pm7zf\") pod \"authentication-operator-69f744f599-n25dz\" (UID: \"fd2f0b92-9f4d-4c2c-b83f-7530706cfa09\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-n25dz" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.069884 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-797bd\" (UniqueName: \"kubernetes.io/projected/30e5171d-473c-4388-977d-95ce80ae1b79-kube-api-access-797bd\") pod \"cluster-image-registry-operator-dc59b4c8b-zhgdz\" (UID: \"30e5171d-473c-4388-977d-95ce80ae1b79\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhgdz" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.070306 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42ba2ce9-d014-4090-b4b2-21a2ae8cea4c-config\") pod \"machine-api-operator-5694c8668f-22f85\" (UID: \"42ba2ce9-d014-4090-b4b2-21a2ae8cea4c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-22f85" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.081301 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd2f0b92-9f4d-4c2c-b83f-7530706cfa09-serving-cert\") pod \"authentication-operator-69f744f599-n25dz\" (UID: \"fd2f0b92-9f4d-4c2c-b83f-7530706cfa09\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-n25dz" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.074598 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.081367 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abcc561e-04a7-4cf3-a5d3-f2937d0f18c0-config\") pod \"kube-apiserver-operator-766d6c64bb-fjmq5\" (UID: \"abcc561e-04a7-4cf3-a5d3-f2937d0f18c0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fjmq5" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.081395 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd2f0b92-9f4d-4c2c-b83f-7530706cfa09-config\") pod \"authentication-operator-69f744f599-n25dz\" (UID: \"fd2f0b92-9f4d-4c2c-b83f-7530706cfa09\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-n25dz" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.074637 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.081456 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c6cb779-d660-45f8-8fb2-5829d42ee030-config\") pod \"route-controller-manager-6576b87f9c-v6thv\" (UID: \"0c6cb779-d660-45f8-8fb2-5829d42ee030\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v6thv" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.081617 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a51266e-326e-4b5c-a25a-fb61bf1f3757-metrics-certs\") pod \"router-default-5444994796-8cq46\" (UID: \"7a51266e-326e-4b5c-a25a-fb61bf1f3757\") " pod="openshift-ingress/router-default-5444994796-8cq46" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.081692 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/81190a62-b874-433f-8e50-f6dfe001f6d2-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-b8czr\" (UID: \"81190a62-b874-433f-8e50-f6dfe001f6d2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.081716 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpghg\" (UniqueName: \"kubernetes.io/projected/81190a62-b874-433f-8e50-f6dfe001f6d2-kube-api-access-lpghg\") pod \"apiserver-7bbb656c7d-b8czr\" (UID: \"81190a62-b874-433f-8e50-f6dfe001f6d2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.081771 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/81190a62-b874-433f-8e50-f6dfe001f6d2-audit-dir\") pod \"apiserver-7bbb656c7d-b8czr\" (UID: \"81190a62-b874-433f-8e50-f6dfe001f6d2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.081799 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f8cbed1-6086-4fb6-bab5-0563bbcb751c-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-966wr\" (UID: \"7f8cbed1-6086-4fb6-bab5-0563bbcb751c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-966wr" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.081845 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/abcc561e-04a7-4cf3-a5d3-f2937d0f18c0-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-fjmq5\" (UID: \"abcc561e-04a7-4cf3-a5d3-f2937d0f18c0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fjmq5" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.081870 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8km7p\" (UniqueName: \"kubernetes.io/projected/7a51266e-326e-4b5c-a25a-fb61bf1f3757-kube-api-access-8km7p\") pod \"router-default-5444994796-8cq46\" (UID: \"7a51266e-326e-4b5c-a25a-fb61bf1f3757\") " pod="openshift-ingress/router-default-5444994796-8cq46" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.081892 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f2da710c-c8cd-4e5f-8ba3-045858050ade-auth-proxy-config\") pod \"machine-approver-56656f9798-t2wdv\" (UID: \"f2da710c-c8cd-4e5f-8ba3-045858050ade\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t2wdv" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.081938 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/30e5171d-473c-4388-977d-95ce80ae1b79-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-zhgdz\" (UID: \"30e5171d-473c-4388-977d-95ce80ae1b79\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhgdz" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.081966 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcz9m\" (UniqueName: \"kubernetes.io/projected/7f8cbed1-6086-4fb6-bab5-0563bbcb751c-kube-api-access-jcz9m\") pod \"kube-storage-version-migrator-operator-b67b599dd-966wr\" (UID: \"7f8cbed1-6086-4fb6-bab5-0563bbcb751c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-966wr" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.082020 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/7a51266e-326e-4b5c-a25a-fb61bf1f3757-default-certificate\") pod \"router-default-5444994796-8cq46\" (UID: \"7a51266e-326e-4b5c-a25a-fb61bf1f3757\") " pod="openshift-ingress/router-default-5444994796-8cq46" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.082045 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhkrn\" (UniqueName: \"kubernetes.io/projected/0c6cb779-d660-45f8-8fb2-5829d42ee030-kube-api-access-hhkrn\") pod \"route-controller-manager-6576b87f9c-v6thv\" (UID: \"0c6cb779-d660-45f8-8fb2-5829d42ee030\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v6thv" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.082101 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1659f4f-bd6e-4d58-a56b-69b14ab57ed0-config\") pod \"console-operator-58897d9998-p7zgl\" (UID: \"a1659f4f-bd6e-4d58-a56b-69b14ab57ed0\") " pod="openshift-console-operator/console-operator-58897d9998-p7zgl" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.082138 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4lss\" (UniqueName: \"kubernetes.io/projected/f2da710c-c8cd-4e5f-8ba3-045858050ade-kube-api-access-j4lss\") pod \"machine-approver-56656f9798-t2wdv\" (UID: \"f2da710c-c8cd-4e5f-8ba3-045858050ade\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t2wdv" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.082183 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6mfk\" (UniqueName: \"kubernetes.io/projected/a1659f4f-bd6e-4d58-a56b-69b14ab57ed0-kube-api-access-x6mfk\") pod \"console-operator-58897d9998-p7zgl\" (UID: \"a1659f4f-bd6e-4d58-a56b-69b14ab57ed0\") " pod="openshift-console-operator/console-operator-58897d9998-p7zgl" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.082254 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2da710c-c8cd-4e5f-8ba3-045858050ade-config\") pod \"machine-approver-56656f9798-t2wdv\" (UID: \"f2da710c-c8cd-4e5f-8ba3-045858050ade\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t2wdv" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.082287 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/81190a62-b874-433f-8e50-f6dfe001f6d2-audit-policies\") pod \"apiserver-7bbb656c7d-b8czr\" (UID: \"81190a62-b874-433f-8e50-f6dfe001f6d2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.082309 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/30e5171d-473c-4388-977d-95ce80ae1b79-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-zhgdz\" (UID: \"30e5171d-473c-4388-977d-95ce80ae1b79\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhgdz" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.082356 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/81190a62-b874-433f-8e50-f6dfe001f6d2-etcd-client\") pod \"apiserver-7bbb656c7d-b8czr\" (UID: \"81190a62-b874-433f-8e50-f6dfe001f6d2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.082377 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/7a51266e-326e-4b5c-a25a-fb61bf1f3757-stats-auth\") pod \"router-default-5444994796-8cq46\" (UID: \"7a51266e-326e-4b5c-a25a-fb61bf1f3757\") " pod="openshift-ingress/router-default-5444994796-8cq46" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.082423 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fd2f0b92-9f4d-4c2c-b83f-7530706cfa09-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-n25dz\" (UID: \"fd2f0b92-9f4d-4c2c-b83f-7530706cfa09\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-n25dz" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.082445 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/abcc561e-04a7-4cf3-a5d3-f2937d0f18c0-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-fjmq5\" (UID: \"abcc561e-04a7-4cf3-a5d3-f2937d0f18c0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fjmq5" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.082466 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a1659f4f-bd6e-4d58-a56b-69b14ab57ed0-trusted-ca\") pod \"console-operator-58897d9998-p7zgl\" (UID: \"a1659f4f-bd6e-4d58-a56b-69b14ab57ed0\") " pod="openshift-console-operator/console-operator-58897d9998-p7zgl" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.082513 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/81190a62-b874-433f-8e50-f6dfe001f6d2-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-b8czr\" (UID: \"81190a62-b874-433f-8e50-f6dfe001f6d2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.082536 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a1659f4f-bd6e-4d58-a56b-69b14ab57ed0-serving-cert\") pod \"console-operator-58897d9998-p7zgl\" (UID: \"a1659f4f-bd6e-4d58-a56b-69b14ab57ed0\") " pod="openshift-console-operator/console-operator-58897d9998-p7zgl" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.082584 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0c6cb779-d660-45f8-8fb2-5829d42ee030-client-ca\") pod \"route-controller-manager-6576b87f9c-v6thv\" (UID: \"0c6cb779-d660-45f8-8fb2-5829d42ee030\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v6thv" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.082709 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81190a62-b874-433f-8e50-f6dfe001f6d2-serving-cert\") pod \"apiserver-7bbb656c7d-b8czr\" (UID: \"81190a62-b874-433f-8e50-f6dfe001f6d2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.082737 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/81190a62-b874-433f-8e50-f6dfe001f6d2-encryption-config\") pod \"apiserver-7bbb656c7d-b8czr\" (UID: \"81190a62-b874-433f-8e50-f6dfe001f6d2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.082740 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/30e5171d-473c-4388-977d-95ce80ae1b79-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-zhgdz\" (UID: \"30e5171d-473c-4388-977d-95ce80ae1b79\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhgdz" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.082812 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/42ba2ce9-d014-4090-b4b2-21a2ae8cea4c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-22f85\" (UID: \"42ba2ce9-d014-4090-b4b2-21a2ae8cea4c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-22f85" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.082905 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0c6cb779-d660-45f8-8fb2-5829d42ee030-serving-cert\") pod \"route-controller-manager-6576b87f9c-v6thv\" (UID: \"0c6cb779-d660-45f8-8fb2-5829d42ee030\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v6thv" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.083335 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2da710c-c8cd-4e5f-8ba3-045858050ade-config\") pod \"machine-approver-56656f9798-t2wdv\" (UID: \"f2da710c-c8cd-4e5f-8ba3-045858050ade\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t2wdv" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.083603 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/81190a62-b874-433f-8e50-f6dfe001f6d2-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-b8czr\" (UID: \"81190a62-b874-433f-8e50-f6dfe001f6d2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.083784 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/81190a62-b874-433f-8e50-f6dfe001f6d2-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-b8czr\" (UID: \"81190a62-b874-433f-8e50-f6dfe001f6d2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.082513 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c6cb779-d660-45f8-8fb2-5829d42ee030-config\") pod \"route-controller-manager-6576b87f9c-v6thv\" (UID: \"0c6cb779-d660-45f8-8fb2-5829d42ee030\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v6thv" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.083823 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/81190a62-b874-433f-8e50-f6dfe001f6d2-audit-dir\") pod \"apiserver-7bbb656c7d-b8czr\" (UID: \"81190a62-b874-433f-8e50-f6dfe001f6d2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.083938 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/81190a62-b874-433f-8e50-f6dfe001f6d2-audit-policies\") pod \"apiserver-7bbb656c7d-b8czr\" (UID: \"81190a62-b874-433f-8e50-f6dfe001f6d2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.070908 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/42ba2ce9-d014-4090-b4b2-21a2ae8cea4c-images\") pod \"machine-api-operator-5694c8668f-22f85\" (UID: \"42ba2ce9-d014-4090-b4b2-21a2ae8cea4c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-22f85" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.084534 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f2da710c-c8cd-4e5f-8ba3-045858050ade-auth-proxy-config\") pod \"machine-approver-56656f9798-t2wdv\" (UID: \"f2da710c-c8cd-4e5f-8ba3-045858050ade\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t2wdv" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.084959 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0c6cb779-d660-45f8-8fb2-5829d42ee030-client-ca\") pod \"route-controller-manager-6576b87f9c-v6thv\" (UID: \"0c6cb779-d660-45f8-8fb2-5829d42ee030\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v6thv" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.086047 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.088295 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jpg48"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.088980 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jpg48" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.090472 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f2da710c-c8cd-4e5f-8ba3-045858050ade-machine-approver-tls\") pod \"machine-approver-56656f9798-t2wdv\" (UID: \"f2da710c-c8cd-4e5f-8ba3-045858050ade\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t2wdv" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.090539 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-ccl77"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.091173 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-ccl77" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.091618 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/81190a62-b874-433f-8e50-f6dfe001f6d2-etcd-client\") pod \"apiserver-7bbb656c7d-b8czr\" (UID: \"81190a62-b874-433f-8e50-f6dfe001f6d2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.092973 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-n25dz"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.093924 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/42ba2ce9-d014-4090-b4b2-21a2ae8cea4c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-22f85\" (UID: \"42ba2ce9-d014-4090-b4b2-21a2ae8cea4c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-22f85" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.093979 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/81190a62-b874-433f-8e50-f6dfe001f6d2-serving-cert\") pod \"apiserver-7bbb656c7d-b8czr\" (UID: \"81190a62-b874-433f-8e50-f6dfe001f6d2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.098830 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523765-frmvz"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.099527 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-66zwg"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.099816 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-rx8f2"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.100169 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-rx8f2" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.100737 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523765-frmvz" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.100748 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-66zwg" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.102342 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l8fcf"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.103039 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j99bk"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.103527 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j99bk" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.103712 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bw75d"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.103787 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l8fcf" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.105319 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-tgwt9"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.106706 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-s95mb"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.107563 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-s95mb" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.108572 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-v59s8"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.109501 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v59s8" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.109710 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-27bbf"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.111206 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-p7zgl"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.112147 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.112398 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-rmc7c"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.112928 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-mcjhx"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.115323 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fjmq5"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.117195 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-4bxkk"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.119131 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-rx8f2"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.120660 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523765-frmvz"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.122315 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhgdz"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.124086 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-z7xdz"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.125525 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-6frkl"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.126917 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xrhcj"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.128354 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gqb6w"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.129690 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bncs"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.131081 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-pclrg"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.132145 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.132792 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bjbsw"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.133936 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-shvc6"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.134906 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-966wr"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.135962 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-dlghr"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.141632 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-66zwg"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.141675 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l8fcf"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.141694 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-ccl77"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.146968 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jpg48"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.149305 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-fhpqt"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.150285 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qrwnx"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.151665 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hv6vt"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.153144 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.153285 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-5795c"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.154128 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-5795c" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.154815 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-7jgjj"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.155777 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-7jgjj" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.156485 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-v59s8"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.157488 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-5795c"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.158733 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j99bk"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.159814 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-7jgjj"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.160995 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-znxvs"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.162265 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-znxvs" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.163155 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-znxvs"] Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.172482 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.183419 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fd2f0b92-9f4d-4c2c-b83f-7530706cfa09-service-ca-bundle\") pod \"authentication-operator-69f744f599-n25dz\" (UID: \"fd2f0b92-9f4d-4c2c-b83f-7530706cfa09\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-n25dz" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.183454 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7a51266e-326e-4b5c-a25a-fb61bf1f3757-service-ca-bundle\") pod \"router-default-5444994796-8cq46\" (UID: \"7a51266e-326e-4b5c-a25a-fb61bf1f3757\") " pod="openshift-ingress/router-default-5444994796-8cq46" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.183479 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f8cbed1-6086-4fb6-bab5-0563bbcb751c-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-966wr\" (UID: \"7f8cbed1-6086-4fb6-bab5-0563bbcb751c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-966wr" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.183496 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-797bd\" (UniqueName: \"kubernetes.io/projected/30e5171d-473c-4388-977d-95ce80ae1b79-kube-api-access-797bd\") pod \"cluster-image-registry-operator-dc59b4c8b-zhgdz\" (UID: \"30e5171d-473c-4388-977d-95ce80ae1b79\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhgdz" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.183511 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pm7zf\" (UniqueName: \"kubernetes.io/projected/fd2f0b92-9f4d-4c2c-b83f-7530706cfa09-kube-api-access-pm7zf\") pod \"authentication-operator-69f744f599-n25dz\" (UID: \"fd2f0b92-9f4d-4c2c-b83f-7530706cfa09\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-n25dz" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.183531 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd2f0b92-9f4d-4c2c-b83f-7530706cfa09-serving-cert\") pod \"authentication-operator-69f744f599-n25dz\" (UID: \"fd2f0b92-9f4d-4c2c-b83f-7530706cfa09\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-n25dz" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.183546 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd2f0b92-9f4d-4c2c-b83f-7530706cfa09-config\") pod \"authentication-operator-69f744f599-n25dz\" (UID: \"fd2f0b92-9f4d-4c2c-b83f-7530706cfa09\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-n25dz" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.183564 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abcc561e-04a7-4cf3-a5d3-f2937d0f18c0-config\") pod \"kube-apiserver-operator-766d6c64bb-fjmq5\" (UID: \"abcc561e-04a7-4cf3-a5d3-f2937d0f18c0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fjmq5" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.183586 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a51266e-326e-4b5c-a25a-fb61bf1f3757-metrics-certs\") pod \"router-default-5444994796-8cq46\" (UID: \"7a51266e-326e-4b5c-a25a-fb61bf1f3757\") " pod="openshift-ingress/router-default-5444994796-8cq46" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.183613 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f8cbed1-6086-4fb6-bab5-0563bbcb751c-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-966wr\" (UID: \"7f8cbed1-6086-4fb6-bab5-0563bbcb751c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-966wr" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.183628 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/abcc561e-04a7-4cf3-a5d3-f2937d0f18c0-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-fjmq5\" (UID: \"abcc561e-04a7-4cf3-a5d3-f2937d0f18c0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fjmq5" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.183645 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/30e5171d-473c-4388-977d-95ce80ae1b79-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-zhgdz\" (UID: \"30e5171d-473c-4388-977d-95ce80ae1b79\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhgdz" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.183662 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8km7p\" (UniqueName: \"kubernetes.io/projected/7a51266e-326e-4b5c-a25a-fb61bf1f3757-kube-api-access-8km7p\") pod \"router-default-5444994796-8cq46\" (UID: \"7a51266e-326e-4b5c-a25a-fb61bf1f3757\") " pod="openshift-ingress/router-default-5444994796-8cq46" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.183679 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcz9m\" (UniqueName: \"kubernetes.io/projected/7f8cbed1-6086-4fb6-bab5-0563bbcb751c-kube-api-access-jcz9m\") pod \"kube-storage-version-migrator-operator-b67b599dd-966wr\" (UID: \"7f8cbed1-6086-4fb6-bab5-0563bbcb751c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-966wr" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.183704 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/7a51266e-326e-4b5c-a25a-fb61bf1f3757-default-certificate\") pod \"router-default-5444994796-8cq46\" (UID: \"7a51266e-326e-4b5c-a25a-fb61bf1f3757\") " pod="openshift-ingress/router-default-5444994796-8cq46" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.183725 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1659f4f-bd6e-4d58-a56b-69b14ab57ed0-config\") pod \"console-operator-58897d9998-p7zgl\" (UID: \"a1659f4f-bd6e-4d58-a56b-69b14ab57ed0\") " pod="openshift-console-operator/console-operator-58897d9998-p7zgl" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.184462 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fd2f0b92-9f4d-4c2c-b83f-7530706cfa09-service-ca-bundle\") pod \"authentication-operator-69f744f599-n25dz\" (UID: \"fd2f0b92-9f4d-4c2c-b83f-7530706cfa09\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-n25dz" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.184487 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6mfk\" (UniqueName: \"kubernetes.io/projected/a1659f4f-bd6e-4d58-a56b-69b14ab57ed0-kube-api-access-x6mfk\") pod \"console-operator-58897d9998-p7zgl\" (UID: \"a1659f4f-bd6e-4d58-a56b-69b14ab57ed0\") " pod="openshift-console-operator/console-operator-58897d9998-p7zgl" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.184518 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/30e5171d-473c-4388-977d-95ce80ae1b79-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-zhgdz\" (UID: \"30e5171d-473c-4388-977d-95ce80ae1b79\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhgdz" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.184489 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abcc561e-04a7-4cf3-a5d3-f2937d0f18c0-config\") pod \"kube-apiserver-operator-766d6c64bb-fjmq5\" (UID: \"abcc561e-04a7-4cf3-a5d3-f2937d0f18c0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fjmq5" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.184540 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/7a51266e-326e-4b5c-a25a-fb61bf1f3757-stats-auth\") pod \"router-default-5444994796-8cq46\" (UID: \"7a51266e-326e-4b5c-a25a-fb61bf1f3757\") " pod="openshift-ingress/router-default-5444994796-8cq46" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.184542 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1659f4f-bd6e-4d58-a56b-69b14ab57ed0-config\") pod \"console-operator-58897d9998-p7zgl\" (UID: \"a1659f4f-bd6e-4d58-a56b-69b14ab57ed0\") " pod="openshift-console-operator/console-operator-58897d9998-p7zgl" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.184521 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd2f0b92-9f4d-4c2c-b83f-7530706cfa09-config\") pod \"authentication-operator-69f744f599-n25dz\" (UID: \"fd2f0b92-9f4d-4c2c-b83f-7530706cfa09\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-n25dz" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.184575 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a1659f4f-bd6e-4d58-a56b-69b14ab57ed0-trusted-ca\") pod \"console-operator-58897d9998-p7zgl\" (UID: \"a1659f4f-bd6e-4d58-a56b-69b14ab57ed0\") " pod="openshift-console-operator/console-operator-58897d9998-p7zgl" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.184612 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fd2f0b92-9f4d-4c2c-b83f-7530706cfa09-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-n25dz\" (UID: \"fd2f0b92-9f4d-4c2c-b83f-7530706cfa09\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-n25dz" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.184631 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/abcc561e-04a7-4cf3-a5d3-f2937d0f18c0-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-fjmq5\" (UID: \"abcc561e-04a7-4cf3-a5d3-f2937d0f18c0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fjmq5" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.184648 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a1659f4f-bd6e-4d58-a56b-69b14ab57ed0-serving-cert\") pod \"console-operator-58897d9998-p7zgl\" (UID: \"a1659f4f-bd6e-4d58-a56b-69b14ab57ed0\") " pod="openshift-console-operator/console-operator-58897d9998-p7zgl" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.184668 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/30e5171d-473c-4388-977d-95ce80ae1b79-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-zhgdz\" (UID: \"30e5171d-473c-4388-977d-95ce80ae1b79\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhgdz" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.185358 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fd2f0b92-9f4d-4c2c-b83f-7530706cfa09-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-n25dz\" (UID: \"fd2f0b92-9f4d-4c2c-b83f-7530706cfa09\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-n25dz" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.185658 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a1659f4f-bd6e-4d58-a56b-69b14ab57ed0-trusted-ca\") pod \"console-operator-58897d9998-p7zgl\" (UID: \"a1659f4f-bd6e-4d58-a56b-69b14ab57ed0\") " pod="openshift-console-operator/console-operator-58897d9998-p7zgl" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.185718 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/30e5171d-473c-4388-977d-95ce80ae1b79-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-zhgdz\" (UID: \"30e5171d-473c-4388-977d-95ce80ae1b79\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhgdz" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.186983 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/abcc561e-04a7-4cf3-a5d3-f2937d0f18c0-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-fjmq5\" (UID: \"abcc561e-04a7-4cf3-a5d3-f2937d0f18c0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fjmq5" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.187330 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a1659f4f-bd6e-4d58-a56b-69b14ab57ed0-serving-cert\") pod \"console-operator-58897d9998-p7zgl\" (UID: \"a1659f4f-bd6e-4d58-a56b-69b14ab57ed0\") " pod="openshift-console-operator/console-operator-58897d9998-p7zgl" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.188142 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/30e5171d-473c-4388-977d-95ce80ae1b79-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-zhgdz\" (UID: \"30e5171d-473c-4388-977d-95ce80ae1b79\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhgdz" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.188417 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd2f0b92-9f4d-4c2c-b83f-7530706cfa09-serving-cert\") pod \"authentication-operator-69f744f599-n25dz\" (UID: \"fd2f0b92-9f4d-4c2c-b83f-7530706cfa09\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-n25dz" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.193342 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.212501 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.231995 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.252894 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.272402 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.276827 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/7a51266e-326e-4b5c-a25a-fb61bf1f3757-default-certificate\") pod \"router-default-5444994796-8cq46\" (UID: \"7a51266e-326e-4b5c-a25a-fb61bf1f3757\") " pod="openshift-ingress/router-default-5444994796-8cq46" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.292640 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.298125 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/7a51266e-326e-4b5c-a25a-fb61bf1f3757-stats-auth\") pod \"router-default-5444994796-8cq46\" (UID: \"7a51266e-326e-4b5c-a25a-fb61bf1f3757\") " pod="openshift-ingress/router-default-5444994796-8cq46" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.312224 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.316760 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a51266e-326e-4b5c-a25a-fb61bf1f3757-metrics-certs\") pod \"router-default-5444994796-8cq46\" (UID: \"7a51266e-326e-4b5c-a25a-fb61bf1f3757\") " pod="openshift-ingress/router-default-5444994796-8cq46" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.333211 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.335093 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7a51266e-326e-4b5c-a25a-fb61bf1f3757-service-ca-bundle\") pod \"router-default-5444994796-8cq46\" (UID: \"7a51266e-326e-4b5c-a25a-fb61bf1f3757\") " pod="openshift-ingress/router-default-5444994796-8cq46" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.352340 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.372135 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.393736 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.413721 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.433035 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.452567 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.473480 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.492655 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.513840 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.532649 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.537112 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7f8cbed1-6086-4fb6-bab5-0563bbcb751c-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-966wr\" (UID: \"7f8cbed1-6086-4fb6-bab5-0563bbcb751c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-966wr" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.553369 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.554854 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f8cbed1-6086-4fb6-bab5-0563bbcb751c-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-966wr\" (UID: \"7f8cbed1-6086-4fb6-bab5-0563bbcb751c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-966wr" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.572526 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.593297 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.612323 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.632312 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.652670 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.678405 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.712720 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.733470 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.753293 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.772365 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.802699 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.812467 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.833111 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.853180 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.871860 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.893321 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.912746 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.932690 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.953200 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.972379 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 18 14:57:22 crc kubenswrapper[4896]: I0218 14:57:22.992688 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.013729 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.033126 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.052496 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.071110 4896 request.go:700] Waited for 1.010984192s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager-operator/secrets?fieldSelector=metadata.name%3Dkube-controller-manager-operator-serving-cert&limit=500&resourceVersion=0 Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.072808 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.092744 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.112650 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.134006 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.152926 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.173160 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.192863 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.213894 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.233087 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.253144 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.290296 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkt2f\" (UniqueName: \"kubernetes.io/projected/42ba2ce9-d014-4090-b4b2-21a2ae8cea4c-kube-api-access-kkt2f\") pod \"machine-api-operator-5694c8668f-22f85\" (UID: \"42ba2ce9-d014-4090-b4b2-21a2ae8cea4c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-22f85" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.312081 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4lss\" (UniqueName: \"kubernetes.io/projected/f2da710c-c8cd-4e5f-8ba3-045858050ade-kube-api-access-j4lss\") pod \"machine-approver-56656f9798-t2wdv\" (UID: \"f2da710c-c8cd-4e5f-8ba3-045858050ade\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t2wdv" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.327749 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpghg\" (UniqueName: \"kubernetes.io/projected/81190a62-b874-433f-8e50-f6dfe001f6d2-kube-api-access-lpghg\") pod \"apiserver-7bbb656c7d-b8czr\" (UID: \"81190a62-b874-433f-8e50-f6dfe001f6d2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.347137 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhkrn\" (UniqueName: \"kubernetes.io/projected/0c6cb779-d660-45f8-8fb2-5829d42ee030-kube-api-access-hhkrn\") pod \"route-controller-manager-6576b87f9c-v6thv\" (UID: \"0c6cb779-d660-45f8-8fb2-5829d42ee030\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v6thv" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.353202 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.372759 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.392651 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.412490 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.422732 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-22f85" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.432884 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.452997 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.457599 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t2wdv" Feb 18 14:57:23 crc kubenswrapper[4896]: W0218 14:57:23.469465 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf2da710c_c8cd_4e5f_8ba3_045858050ade.slice/crio-84807791a839fd7b92adf5fa7b79417d5bb249082667b8ac2ae78e28c0322d11 WatchSource:0}: Error finding container 84807791a839fd7b92adf5fa7b79417d5bb249082667b8ac2ae78e28c0322d11: Status 404 returned error can't find the container with id 84807791a839fd7b92adf5fa7b79417d5bb249082667b8ac2ae78e28c0322d11 Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.472728 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v6thv" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.472955 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.513574 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.528345 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.573817 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.573926 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.573982 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.586534 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t2wdv" event={"ID":"f2da710c-c8cd-4e5f-8ba3-045858050ade","Type":"ContainerStarted","Data":"84807791a839fd7b92adf5fa7b79417d5bb249082667b8ac2ae78e28c0322d11"} Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.593053 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.613089 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.633422 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.653997 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.671795 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-22f85"] Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.673947 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.692848 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.695418 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-v6thv"] Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.713171 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.719221 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr"] Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.732804 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.753412 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.778534 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.793213 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.813018 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.833676 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.853677 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.872459 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.892778 4896 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.914914 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.932069 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.952334 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.972942 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 18 14:57:23 crc kubenswrapper[4896]: I0218 14:57:23.993089 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.012499 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.049080 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pm7zf\" (UniqueName: \"kubernetes.io/projected/fd2f0b92-9f4d-4c2c-b83f-7530706cfa09-kube-api-access-pm7zf\") pod \"authentication-operator-69f744f599-n25dz\" (UID: \"fd2f0b92-9f4d-4c2c-b83f-7530706cfa09\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-n25dz" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.070500 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-797bd\" (UniqueName: \"kubernetes.io/projected/30e5171d-473c-4388-977d-95ce80ae1b79-kube-api-access-797bd\") pod \"cluster-image-registry-operator-dc59b4c8b-zhgdz\" (UID: \"30e5171d-473c-4388-977d-95ce80ae1b79\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhgdz" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.071923 4896 request.go:700] Waited for 1.888131658s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/serviceaccounts/cluster-image-registry-operator/token Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.089954 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/30e5171d-473c-4388-977d-95ce80ae1b79-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-zhgdz\" (UID: \"30e5171d-473c-4388-977d-95ce80ae1b79\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhgdz" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.110763 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/abcc561e-04a7-4cf3-a5d3-f2937d0f18c0-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-fjmq5\" (UID: \"abcc561e-04a7-4cf3-a5d3-f2937d0f18c0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fjmq5" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.136257 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcz9m\" (UniqueName: \"kubernetes.io/projected/7f8cbed1-6086-4fb6-bab5-0563bbcb751c-kube-api-access-jcz9m\") pod \"kube-storage-version-migrator-operator-b67b599dd-966wr\" (UID: \"7f8cbed1-6086-4fb6-bab5-0563bbcb751c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-966wr" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.145985 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8km7p\" (UniqueName: \"kubernetes.io/projected/7a51266e-326e-4b5c-a25a-fb61bf1f3757-kube-api-access-8km7p\") pod \"router-default-5444994796-8cq46\" (UID: \"7a51266e-326e-4b5c-a25a-fb61bf1f3757\") " pod="openshift-ingress/router-default-5444994796-8cq46" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.177534 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6mfk\" (UniqueName: \"kubernetes.io/projected/a1659f4f-bd6e-4d58-a56b-69b14ab57ed0-kube-api-access-x6mfk\") pod \"console-operator-58897d9998-p7zgl\" (UID: \"a1659f4f-bd6e-4d58-a56b-69b14ab57ed0\") " pod="openshift-console-operator/console-operator-58897d9998-p7zgl" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.207149 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-272j4\" (UniqueName: \"kubernetes.io/projected/31b0687b-5c42-4259-b96d-284c1aa487b4-kube-api-access-272j4\") pod \"controller-manager-879f6c89f-27bbf\" (UID: \"31b0687b-5c42-4259-b96d-284c1aa487b4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-27bbf" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.207188 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/37bf7cb8-e127-458a-ab07-efb2c111ce02-bound-sa-token\") pod \"ingress-operator-5b745b69d9-dlghr\" (UID: \"37bf7cb8-e127-458a-ab07-efb2c111ce02\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dlghr" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.207331 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f8abcf75-9577-4dfe-a252-40aff339ffce-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-gqb6w\" (UID: \"f8abcf75-9577-4dfe-a252-40aff339ffce\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gqb6w" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.207377 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.207435 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71145266-38d8-495b-83c2-cff3c68e705d-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2ftds\" (UID: \"71145266-38d8-495b-83c2-cff3c68e705d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ftds" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.207461 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/710fc779-629b-4d37-a820-e08f5289677f-audit\") pod \"apiserver-76f77b778f-tgwt9\" (UID: \"710fc779-629b-4d37-a820-e08f5289677f\") " pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.207513 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/37bf7cb8-e127-458a-ab07-efb2c111ce02-trusted-ca\") pod \"ingress-operator-5b745b69d9-dlghr\" (UID: \"37bf7cb8-e127-458a-ab07-efb2c111ce02\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dlghr" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.207533 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/76cb1e0e-4321-4450-9fcb-b505a7a1d969-audit-dir\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.207554 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/21fb3328-3882-4644-9a94-b7949f48db54-auth-proxy-config\") pod \"machine-config-operator-74547568cd-6frkl\" (UID: \"21fb3328-3882-4644-9a94-b7949f48db54\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6frkl" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.207590 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31b0687b-5c42-4259-b96d-284c1aa487b4-serving-cert\") pod \"controller-manager-879f6c89f-27bbf\" (UID: \"31b0687b-5c42-4259-b96d-284c1aa487b4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-27bbf" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.207627 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/21fb3328-3882-4644-9a94-b7949f48db54-proxy-tls\") pod \"machine-config-operator-74547568cd-6frkl\" (UID: \"21fb3328-3882-4644-9a94-b7949f48db54\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6frkl" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.207655 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/351aa26e-939a-4b69-a7ee-55be7111ee6e-ca-trust-extracted\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.207675 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/710fc779-629b-4d37-a820-e08f5289677f-config\") pod \"apiserver-76f77b778f-tgwt9\" (UID: \"710fc779-629b-4d37-a820-e08f5289677f\") " pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.207707 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/918ef2a1-8828-4b17-b30a-de957f009104-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-z7xdz\" (UID: \"918ef2a1-8828-4b17-b30a-de957f009104\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-z7xdz" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.207728 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37bf7cb8-e127-458a-ab07-efb2c111ce02-metrics-tls\") pod \"ingress-operator-5b745b69d9-dlghr\" (UID: \"37bf7cb8-e127-458a-ab07-efb2c111ce02\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dlghr" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.207751 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.207774 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.207799 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/fa6c59b4-6823-47ef-a8a6-52c94e53ef4c-console-config\") pod \"console-f9d7485db-4bxkk\" (UID: \"fa6c59b4-6823-47ef-a8a6-52c94e53ef4c\") " pod="openshift-console/console-f9d7485db-4bxkk" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.207909 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/fa6c59b4-6823-47ef-a8a6-52c94e53ef4c-oauth-serving-cert\") pod \"console-f9d7485db-4bxkk\" (UID: \"fa6c59b4-6823-47ef-a8a6-52c94e53ef4c\") " pod="openshift-console/console-f9d7485db-4bxkk" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.207954 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxhp2\" (UniqueName: \"kubernetes.io/projected/f8abcf75-9577-4dfe-a252-40aff339ffce-kube-api-access-gxhp2\") pod \"openshift-controller-manager-operator-756b6f6bc6-gqb6w\" (UID: \"f8abcf75-9577-4dfe-a252-40aff339ffce\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gqb6w" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.207980 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f88rx\" (UniqueName: \"kubernetes.io/projected/351aa26e-939a-4b69-a7ee-55be7111ee6e-kube-api-access-f88rx\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.208035 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdlmm\" (UniqueName: \"kubernetes.io/projected/710fc779-629b-4d37-a820-e08f5289677f-kube-api-access-mdlmm\") pod \"apiserver-76f77b778f-tgwt9\" (UID: \"710fc779-629b-4d37-a820-e08f5289677f\") " pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.208140 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/fa6c59b4-6823-47ef-a8a6-52c94e53ef4c-console-serving-cert\") pod \"console-f9d7485db-4bxkk\" (UID: \"fa6c59b4-6823-47ef-a8a6-52c94e53ef4c\") " pod="openshift-console/console-f9d7485db-4bxkk" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.208260 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23b10824-dc70-459f-b037-6f8f00ae6a04-serving-cert\") pod \"openshift-config-operator-7777fb866f-mcjhx\" (UID: \"23b10824-dc70-459f-b037-6f8f00ae6a04\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mcjhx" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.208385 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/351aa26e-939a-4b69-a7ee-55be7111ee6e-installation-pull-secrets\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.208481 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/351aa26e-939a-4b69-a7ee-55be7111ee6e-registry-tls\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.208521 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/351aa26e-939a-4b69-a7ee-55be7111ee6e-registry-certificates\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.208554 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bjgv\" (UniqueName: \"kubernetes.io/projected/036e7602-7608-4033-b18d-d859a7eddebd-kube-api-access-4bjgv\") pod \"downloads-7954f5f757-6v4cn\" (UID: \"036e7602-7608-4033-b18d-d859a7eddebd\") " pod="openshift-console/downloads-7954f5f757-6v4cn" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.208599 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/76cb1e0e-4321-4450-9fcb-b505a7a1d969-audit-policies\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.208657 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/710fc779-629b-4d37-a820-e08f5289677f-trusted-ca-bundle\") pod \"apiserver-76f77b778f-tgwt9\" (UID: \"710fc779-629b-4d37-a820-e08f5289677f\") " pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.208687 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/710fc779-629b-4d37-a820-e08f5289677f-node-pullsecrets\") pod \"apiserver-76f77b778f-tgwt9\" (UID: \"710fc779-629b-4d37-a820-e08f5289677f\") " pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.208920 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/fa6c59b4-6823-47ef-a8a6-52c94e53ef4c-console-oauth-config\") pod \"console-f9d7485db-4bxkk\" (UID: \"fa6c59b4-6823-47ef-a8a6-52c94e53ef4c\") " pod="openshift-console/console-f9d7485db-4bxkk" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.208946 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31b0687b-5c42-4259-b96d-284c1aa487b4-config\") pod \"controller-manager-879f6c89f-27bbf\" (UID: \"31b0687b-5c42-4259-b96d-284c1aa487b4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-27bbf" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.209227 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/23b10824-dc70-459f-b037-6f8f00ae6a04-available-featuregates\") pod \"openshift-config-operator-7777fb866f-mcjhx\" (UID: \"23b10824-dc70-459f-b037-6f8f00ae6a04\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mcjhx" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.209281 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6pk7\" (UniqueName: \"kubernetes.io/projected/76cb1e0e-4321-4450-9fcb-b505a7a1d969-kube-api-access-q6pk7\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.209297 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.209385 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r28rq\" (UniqueName: \"kubernetes.io/projected/3cd5b209-b64c-4ce1-85e2-e00e568b4bbe-kube-api-access-r28rq\") pod \"cluster-samples-operator-665b6dd947-bw75d\" (UID: \"3cd5b209-b64c-4ce1-85e2-e00e568b4bbe\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bw75d" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.209403 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/710fc779-629b-4d37-a820-e08f5289677f-image-import-ca\") pod \"apiserver-76f77b778f-tgwt9\" (UID: \"710fc779-629b-4d37-a820-e08f5289677f\") " pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.209542 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlg8k\" (UniqueName: \"kubernetes.io/projected/fa6c59b4-6823-47ef-a8a6-52c94e53ef4c-kube-api-access-wlg8k\") pod \"console-f9d7485db-4bxkk\" (UID: \"fa6c59b4-6823-47ef-a8a6-52c94e53ef4c\") " pod="openshift-console/console-f9d7485db-4bxkk" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.209585 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/17344065-a988-48bb-aff9-d51ffd76f54b-metrics-tls\") pod \"dns-operator-744455d44c-rmc7c\" (UID: \"17344065-a988-48bb-aff9-d51ffd76f54b\") " pod="openshift-dns-operator/dns-operator-744455d44c-rmc7c" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.209610 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.209659 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.209680 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqzfn\" (UniqueName: \"kubernetes.io/projected/17344065-a988-48bb-aff9-d51ffd76f54b-kube-api-access-cqzfn\") pod \"dns-operator-744455d44c-rmc7c\" (UID: \"17344065-a988-48bb-aff9-d51ffd76f54b\") " pod="openshift-dns-operator/dns-operator-744455d44c-rmc7c" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.209728 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tb9ks\" (UniqueName: \"kubernetes.io/projected/918ef2a1-8828-4b17-b30a-de957f009104-kube-api-access-tb9ks\") pod \"multus-admission-controller-857f4d67dd-z7xdz\" (UID: \"918ef2a1-8828-4b17-b30a-de957f009104\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-z7xdz" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.209766 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/710fc779-629b-4d37-a820-e08f5289677f-serving-cert\") pod \"apiserver-76f77b778f-tgwt9\" (UID: \"710fc779-629b-4d37-a820-e08f5289677f\") " pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.209789 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/710fc779-629b-4d37-a820-e08f5289677f-encryption-config\") pod \"apiserver-76f77b778f-tgwt9\" (UID: \"710fc779-629b-4d37-a820-e08f5289677f\") " pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.209811 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6pnr\" (UniqueName: \"kubernetes.io/projected/21fb3328-3882-4644-9a94-b7949f48db54-kube-api-access-j6pnr\") pod \"machine-config-operator-74547568cd-6frkl\" (UID: \"21fb3328-3882-4644-9a94-b7949f48db54\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6frkl" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.209837 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75xt2\" (UniqueName: \"kubernetes.io/projected/71145266-38d8-495b-83c2-cff3c68e705d-kube-api-access-75xt2\") pod \"openshift-apiserver-operator-796bbdcf4f-2ftds\" (UID: \"71145266-38d8-495b-83c2-cff3c68e705d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ftds" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.209861 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/710fc779-629b-4d37-a820-e08f5289677f-etcd-serving-ca\") pod \"apiserver-76f77b778f-tgwt9\" (UID: \"710fc779-629b-4d37-a820-e08f5289677f\") " pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.209912 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfwkt\" (UniqueName: \"kubernetes.io/projected/23b10824-dc70-459f-b037-6f8f00ae6a04-kube-api-access-cfwkt\") pod \"openshift-config-operator-7777fb866f-mcjhx\" (UID: \"23b10824-dc70-459f-b037-6f8f00ae6a04\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mcjhx" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.209935 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.209996 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/351aa26e-939a-4b69-a7ee-55be7111ee6e-trusted-ca\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.210032 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/351aa26e-939a-4b69-a7ee-55be7111ee6e-bound-sa-token\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.210059 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71145266-38d8-495b-83c2-cff3c68e705d-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2ftds\" (UID: \"71145266-38d8-495b-83c2-cff3c68e705d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ftds" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.210083 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dl5w6\" (UniqueName: \"kubernetes.io/projected/37bf7cb8-e127-458a-ab07-efb2c111ce02-kube-api-access-dl5w6\") pod \"ingress-operator-5b745b69d9-dlghr\" (UID: \"37bf7cb8-e127-458a-ab07-efb2c111ce02\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dlghr" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.210112 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/710fc779-629b-4d37-a820-e08f5289677f-audit-dir\") pod \"apiserver-76f77b778f-tgwt9\" (UID: \"710fc779-629b-4d37-a820-e08f5289677f\") " pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.210136 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fa6c59b4-6823-47ef-a8a6-52c94e53ef4c-trusted-ca-bundle\") pod \"console-f9d7485db-4bxkk\" (UID: \"fa6c59b4-6823-47ef-a8a6-52c94e53ef4c\") " pod="openshift-console/console-f9d7485db-4bxkk" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.210183 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/710fc779-629b-4d37-a820-e08f5289677f-etcd-client\") pod \"apiserver-76f77b778f-tgwt9\" (UID: \"710fc779-629b-4d37-a820-e08f5289677f\") " pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.210213 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fa6c59b4-6823-47ef-a8a6-52c94e53ef4c-service-ca\") pod \"console-f9d7485db-4bxkk\" (UID: \"fa6c59b4-6823-47ef-a8a6-52c94e53ef4c\") " pod="openshift-console/console-f9d7485db-4bxkk" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.210256 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.210285 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/31b0687b-5c42-4259-b96d-284c1aa487b4-client-ca\") pod \"controller-manager-879f6c89f-27bbf\" (UID: \"31b0687b-5c42-4259-b96d-284c1aa487b4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-27bbf" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.210304 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.210326 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/3cd5b209-b64c-4ce1-85e2-e00e568b4bbe-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-bw75d\" (UID: \"3cd5b209-b64c-4ce1-85e2-e00e568b4bbe\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bw75d" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.210363 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.210385 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8abcf75-9577-4dfe-a252-40aff339ffce-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-gqb6w\" (UID: \"f8abcf75-9577-4dfe-a252-40aff339ffce\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gqb6w" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.210521 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/21fb3328-3882-4644-9a94-b7949f48db54-images\") pod \"machine-config-operator-74547568cd-6frkl\" (UID: \"21fb3328-3882-4644-9a94-b7949f48db54\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6frkl" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.210571 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/31b0687b-5c42-4259-b96d-284c1aa487b4-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-27bbf\" (UID: \"31b0687b-5c42-4259-b96d-284c1aa487b4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-27bbf" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.210590 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.210614 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: E0218 14:57:24.210667 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:24.710654247 +0000 UTC m=+143.639988460 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.231779 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-n25dz" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.260720 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-p7zgl" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.279714 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhgdz" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.312137 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.312374 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfwkt\" (UniqueName: \"kubernetes.io/projected/23b10824-dc70-459f-b037-6f8f00ae6a04-kube-api-access-cfwkt\") pod \"openshift-config-operator-7777fb866f-mcjhx\" (UID: \"23b10824-dc70-459f-b037-6f8f00ae6a04\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mcjhx" Feb 18 14:57:24 crc kubenswrapper[4896]: E0218 14:57:24.312590 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:24.812562346 +0000 UTC m=+143.741896559 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.312903 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/710fc779-629b-4d37-a820-e08f5289677f-etcd-serving-ca\") pod \"apiserver-76f77b778f-tgwt9\" (UID: \"710fc779-629b-4d37-a820-e08f5289677f\") " pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.312984 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/351aa26e-939a-4b69-a7ee-55be7111ee6e-trusted-ca\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.313008 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dl5w6\" (UniqueName: \"kubernetes.io/projected/37bf7cb8-e127-458a-ab07-efb2c111ce02-kube-api-access-dl5w6\") pod \"ingress-operator-5b745b69d9-dlghr\" (UID: \"37bf7cb8-e127-458a-ab07-efb2c111ce02\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dlghr" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.313076 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fa6c59b4-6823-47ef-a8a6-52c94e53ef4c-trusted-ca-bundle\") pod \"console-f9d7485db-4bxkk\" (UID: \"fa6c59b4-6823-47ef-a8a6-52c94e53ef4c\") " pod="openshift-console/console-f9d7485db-4bxkk" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.313291 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/351aa26e-939a-4b69-a7ee-55be7111ee6e-bound-sa-token\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.313318 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/710fc779-629b-4d37-a820-e08f5289677f-etcd-client\") pod \"apiserver-76f77b778f-tgwt9\" (UID: \"710fc779-629b-4d37-a820-e08f5289677f\") " pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.315405 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.315437 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/dcf21347-4c2f-4d66-9137-c24ee63a31de-mountpoint-dir\") pod \"csi-hostpathplugin-7jgjj\" (UID: \"dcf21347-4c2f-4d66-9137-c24ee63a31de\") " pod="hostpath-provisioner/csi-hostpathplugin-7jgjj" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.315486 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b60354fe-e785-4e88-af1c-0d846d9ff59a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bjbsw\" (UID: \"b60354fe-e785-4e88-af1c-0d846d9ff59a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bjbsw" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.315542 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8abcf75-9577-4dfe-a252-40aff339ffce-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-gqb6w\" (UID: \"f8abcf75-9577-4dfe-a252-40aff339ffce\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gqb6w" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.315569 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0f95dce4-6d68-44ac-9dd3-215664bb1d75-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xrhcj\" (UID: \"0f95dce4-6d68-44ac-9dd3-215664bb1d75\") " pod="openshift-marketplace/marketplace-operator-79b997595-xrhcj" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.315595 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dd10c491-df5d-43a1-961d-613e2a90be23-config-volume\") pod \"collect-profiles-29523765-frmvz\" (UID: \"dd10c491-df5d-43a1-961d-613e2a90be23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523765-frmvz" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.315618 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dd10c491-df5d-43a1-961d-613e2a90be23-secret-volume\") pod \"collect-profiles-29523765-frmvz\" (UID: \"dd10c491-df5d-43a1-961d-613e2a90be23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523765-frmvz" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.315659 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2573759f-2db1-452a-b310-f2a170099ec6-serving-cert\") pod \"service-ca-operator-777779d784-rx8f2\" (UID: \"2573759f-2db1-452a-b310-f2a170099ec6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-rx8f2" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.315746 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/31b0687b-5c42-4259-b96d-284c1aa487b4-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-27bbf\" (UID: \"31b0687b-5c42-4259-b96d-284c1aa487b4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-27bbf" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.315828 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fjmq5" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.315911 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fa6c59b4-6823-47ef-a8a6-52c94e53ef4c-trusted-ca-bundle\") pod \"console-f9d7485db-4bxkk\" (UID: \"fa6c59b4-6823-47ef-a8a6-52c94e53ef4c\") " pod="openshift-console/console-f9d7485db-4bxkk" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.315964 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/37bf7cb8-e127-458a-ab07-efb2c111ce02-bound-sa-token\") pod \"ingress-operator-5b745b69d9-dlghr\" (UID: \"37bf7cb8-e127-458a-ab07-efb2c111ce02\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dlghr" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.316736 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8abcf75-9577-4dfe-a252-40aff339ffce-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-gqb6w\" (UID: \"f8abcf75-9577-4dfe-a252-40aff339ffce\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gqb6w" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.316776 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f8abcf75-9577-4dfe-a252-40aff339ffce-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-gqb6w\" (UID: \"f8abcf75-9577-4dfe-a252-40aff339ffce\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gqb6w" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.316841 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0f95dce4-6d68-44ac-9dd3-215664bb1d75-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xrhcj\" (UID: \"0f95dce4-6d68-44ac-9dd3-215664bb1d75\") " pod="openshift-marketplace/marketplace-operator-79b997595-xrhcj" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.317039 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71145266-38d8-495b-83c2-cff3c68e705d-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2ftds\" (UID: \"71145266-38d8-495b-83c2-cff3c68e705d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ftds" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.317076 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.317110 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/8db0982f-0e1d-493c-8b32-64190372fcb9-signing-cabundle\") pod \"service-ca-9c57cc56f-ccl77\" (UID: \"8db0982f-0e1d-493c-8b32-64190372fcb9\") " pod="openshift-service-ca/service-ca-9c57cc56f-ccl77" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.317145 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/013d2ce9-60ac-4abf-945d-3340ffa55b72-node-bootstrap-token\") pod \"machine-config-server-s95mb\" (UID: \"013d2ce9-60ac-4abf-945d-3340ffa55b72\") " pod="openshift-machine-config-operator/machine-config-server-s95mb" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.317162 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/dcf21347-4c2f-4d66-9137-c24ee63a31de-csi-data-dir\") pod \"csi-hostpathplugin-7jgjj\" (UID: \"dcf21347-4c2f-4d66-9137-c24ee63a31de\") " pod="hostpath-provisioner/csi-hostpathplugin-7jgjj" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.317306 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/31b0687b-5c42-4259-b96d-284c1aa487b4-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-27bbf\" (UID: \"31b0687b-5c42-4259-b96d-284c1aa487b4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-27bbf" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.317421 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/37bf7cb8-e127-458a-ab07-efb2c111ce02-trusted-ca\") pod \"ingress-operator-5b745b69d9-dlghr\" (UID: \"37bf7cb8-e127-458a-ab07-efb2c111ce02\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dlghr" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.317447 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/76cb1e0e-4321-4450-9fcb-b505a7a1d969-audit-dir\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.317491 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/21fb3328-3882-4644-9a94-b7949f48db54-auth-proxy-config\") pod \"machine-config-operator-74547568cd-6frkl\" (UID: \"21fb3328-3882-4644-9a94-b7949f48db54\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6frkl" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.317523 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/7b9d6a43-e957-4d02-9928-020f9f17f7a9-tmpfs\") pod \"packageserver-d55dfcdfc-l8fcf\" (UID: \"7b9d6a43-e957-4d02-9928-020f9f17f7a9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l8fcf" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.317533 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/351aa26e-939a-4b69-a7ee-55be7111ee6e-trusted-ca\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.317717 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/21fb3328-3882-4644-9a94-b7949f48db54-proxy-tls\") pod \"machine-config-operator-74547568cd-6frkl\" (UID: \"21fb3328-3882-4644-9a94-b7949f48db54\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6frkl" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.318269 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/21fb3328-3882-4644-9a94-b7949f48db54-auth-proxy-config\") pod \"machine-config-operator-74547568cd-6frkl\" (UID: \"21fb3328-3882-4644-9a94-b7949f48db54\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6frkl" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.317720 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71145266-38d8-495b-83c2-cff3c68e705d-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2ftds\" (UID: \"71145266-38d8-495b-83c2-cff3c68e705d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ftds" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.318386 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/918ef2a1-8828-4b17-b30a-de957f009104-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-z7xdz\" (UID: \"918ef2a1-8828-4b17-b30a-de957f009104\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-z7xdz" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.318449 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/fa6c59b4-6823-47ef-a8a6-52c94e53ef4c-console-config\") pod \"console-f9d7485db-4bxkk\" (UID: \"fa6c59b4-6823-47ef-a8a6-52c94e53ef4c\") " pod="openshift-console/console-f9d7485db-4bxkk" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.318479 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/fa6c59b4-6823-47ef-a8a6-52c94e53ef4c-oauth-serving-cert\") pod \"console-f9d7485db-4bxkk\" (UID: \"fa6c59b4-6823-47ef-a8a6-52c94e53ef4c\") " pod="openshift-console/console-f9d7485db-4bxkk" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.318501 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxhp2\" (UniqueName: \"kubernetes.io/projected/f8abcf75-9577-4dfe-a252-40aff339ffce-kube-api-access-gxhp2\") pod \"openshift-controller-manager-operator-756b6f6bc6-gqb6w\" (UID: \"f8abcf75-9577-4dfe-a252-40aff339ffce\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gqb6w" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.318531 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4122c286-47e3-460d-bd25-323012847f12-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-2bncs\" (UID: \"4122c286-47e3-460d-bd25-323012847f12\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bncs" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.318602 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.318629 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdlmm\" (UniqueName: \"kubernetes.io/projected/710fc779-629b-4d37-a820-e08f5289677f-kube-api-access-mdlmm\") pod \"apiserver-76f77b778f-tgwt9\" (UID: \"710fc779-629b-4d37-a820-e08f5289677f\") " pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.318653 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/fa6c59b4-6823-47ef-a8a6-52c94e53ef4c-console-serving-cert\") pod \"console-f9d7485db-4bxkk\" (UID: \"fa6c59b4-6823-47ef-a8a6-52c94e53ef4c\") " pod="openshift-console/console-f9d7485db-4bxkk" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.318961 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/710fc779-629b-4d37-a820-e08f5289677f-etcd-serving-ca\") pod \"apiserver-76f77b778f-tgwt9\" (UID: \"710fc779-629b-4d37-a820-e08f5289677f\") " pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.319202 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/76cb1e0e-4321-4450-9fcb-b505a7a1d969-audit-dir\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.319391 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/fa6c59b4-6823-47ef-a8a6-52c94e53ef4c-console-config\") pod \"console-f9d7485db-4bxkk\" (UID: \"fa6c59b4-6823-47ef-a8a6-52c94e53ef4c\") " pod="openshift-console/console-f9d7485db-4bxkk" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.320128 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/fa6c59b4-6823-47ef-a8a6-52c94e53ef4c-oauth-serving-cert\") pod \"console-f9d7485db-4bxkk\" (UID: \"fa6c59b4-6823-47ef-a8a6-52c94e53ef4c\") " pod="openshift-console/console-f9d7485db-4bxkk" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.320836 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.321049 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/37bf7cb8-e127-458a-ab07-efb2c111ce02-trusted-ca\") pod \"ingress-operator-5b745b69d9-dlghr\" (UID: \"37bf7cb8-e127-458a-ab07-efb2c111ce02\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dlghr" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.321364 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5c0e215f-d940-415a-90d3-f5faf3e537de-profile-collector-cert\") pod \"olm-operator-6b444d44fb-j99bk\" (UID: \"5c0e215f-d940-415a-90d3-f5faf3e537de\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j99bk" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.322281 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6225d3ae-2a4e-479c-aa70-6fdbd2ee488b-serving-cert\") pod \"etcd-operator-b45778765-fhpqt\" (UID: \"6225d3ae-2a4e-479c-aa70-6fdbd2ee488b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fhpqt" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.322336 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jnkn\" (UniqueName: \"kubernetes.io/projected/b0388f1f-8b9c-46e4-a7f0-1fad187d6d65-kube-api-access-5jnkn\") pod \"machine-config-controller-84d6567774-v59s8\" (UID: \"b0388f1f-8b9c-46e4-a7f0-1fad187d6d65\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v59s8" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.323808 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324033 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/21fb3328-3882-4644-9a94-b7949f48db54-proxy-tls\") pod \"machine-config-operator-74547568cd-6frkl\" (UID: \"21fb3328-3882-4644-9a94-b7949f48db54\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6frkl" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324158 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgmr5\" (UniqueName: \"kubernetes.io/projected/0e9fc2ab-6d3d-441d-8591-3d64ce77c62d-kube-api-access-pgmr5\") pod \"migrator-59844c95c7-pclrg\" (UID: \"0e9fc2ab-6d3d-441d-8591-3d64ce77c62d\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pclrg" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324209 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4122c286-47e3-460d-bd25-323012847f12-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-2bncs\" (UID: \"4122c286-47e3-460d-bd25-323012847f12\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bncs" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324249 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/710fc779-629b-4d37-a820-e08f5289677f-trusted-ca-bundle\") pod \"apiserver-76f77b778f-tgwt9\" (UID: \"710fc779-629b-4d37-a820-e08f5289677f\") " pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324279 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8fa6844e-65a5-401e-b329-2a49dc5debe4-cert\") pod \"ingress-canary-znxvs\" (UID: \"8fa6844e-65a5-401e-b329-2a49dc5debe4\") " pod="openshift-ingress-canary/ingress-canary-znxvs" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324301 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31b0687b-5c42-4259-b96d-284c1aa487b4-config\") pod \"controller-manager-879f6c89f-27bbf\" (UID: \"31b0687b-5c42-4259-b96d-284c1aa487b4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-27bbf" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324319 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/23b10824-dc70-459f-b037-6f8f00ae6a04-available-featuregates\") pod \"openshift-config-operator-7777fb866f-mcjhx\" (UID: \"23b10824-dc70-459f-b037-6f8f00ae6a04\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mcjhx" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324336 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6pk7\" (UniqueName: \"kubernetes.io/projected/76cb1e0e-4321-4450-9fcb-b505a7a1d969-kube-api-access-q6pk7\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324353 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/013d2ce9-60ac-4abf-945d-3340ffa55b72-certs\") pod \"machine-config-server-s95mb\" (UID: \"013d2ce9-60ac-4abf-945d-3340ffa55b72\") " pod="openshift-machine-config-operator/machine-config-server-s95mb" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324373 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/dcf21347-4c2f-4d66-9137-c24ee63a31de-plugins-dir\") pod \"csi-hostpathplugin-7jgjj\" (UID: \"dcf21347-4c2f-4d66-9137-c24ee63a31de\") " pod="hostpath-provisioner/csi-hostpathplugin-7jgjj" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324393 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r28rq\" (UniqueName: \"kubernetes.io/projected/3cd5b209-b64c-4ce1-85e2-e00e568b4bbe-kube-api-access-r28rq\") pod \"cluster-samples-operator-665b6dd947-bw75d\" (UID: \"3cd5b209-b64c-4ce1-85e2-e00e568b4bbe\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bw75d" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324411 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b0388f1f-8b9c-46e4-a7f0-1fad187d6d65-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-v59s8\" (UID: \"b0388f1f-8b9c-46e4-a7f0-1fad187d6d65\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v59s8" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324424 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324437 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b0388f1f-8b9c-46e4-a7f0-1fad187d6d65-proxy-tls\") pod \"machine-config-controller-84d6567774-v59s8\" (UID: \"b0388f1f-8b9c-46e4-a7f0-1fad187d6d65\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v59s8" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324456 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6pnr\" (UniqueName: \"kubernetes.io/projected/21fb3328-3882-4644-9a94-b7949f48db54-kube-api-access-j6pnr\") pod \"machine-config-operator-74547568cd-6frkl\" (UID: \"21fb3328-3882-4644-9a94-b7949f48db54\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6frkl" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324476 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q5kn\" (UniqueName: \"kubernetes.io/projected/2573759f-2db1-452a-b310-f2a170099ec6-kube-api-access-8q5kn\") pod \"service-ca-operator-777779d784-rx8f2\" (UID: \"2573759f-2db1-452a-b310-f2a170099ec6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-rx8f2" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324493 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tb9ks\" (UniqueName: \"kubernetes.io/projected/918ef2a1-8828-4b17-b30a-de957f009104-kube-api-access-tb9ks\") pod \"multus-admission-controller-857f4d67dd-z7xdz\" (UID: \"918ef2a1-8828-4b17-b30a-de957f009104\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-z7xdz" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324522 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75xt2\" (UniqueName: \"kubernetes.io/projected/71145266-38d8-495b-83c2-cff3c68e705d-kube-api-access-75xt2\") pod \"openshift-apiserver-operator-796bbdcf4f-2ftds\" (UID: \"71145266-38d8-495b-83c2-cff3c68e705d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ftds" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324541 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324559 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/64e32b45-d067-4151-980d-36ea8f3116e9-profile-collector-cert\") pod \"catalog-operator-68c6474976-66zwg\" (UID: \"64e32b45-d067-4151-980d-36ea8f3116e9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-66zwg" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324582 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2573759f-2db1-452a-b310-f2a170099ec6-config\") pod \"service-ca-operator-777779d784-rx8f2\" (UID: \"2573759f-2db1-452a-b310-f2a170099ec6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-rx8f2" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324607 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rbcf\" (UniqueName: \"kubernetes.io/projected/86b5fc8f-1b76-43ad-b283-0d46697be7aa-kube-api-access-5rbcf\") pod \"package-server-manager-789f6589d5-hv6vt\" (UID: \"86b5fc8f-1b76-43ad-b283-0d46697be7aa\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hv6vt" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324602 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/fa6c59b4-6823-47ef-a8a6-52c94e53ef4c-console-serving-cert\") pod \"console-f9d7485db-4bxkk\" (UID: \"fa6c59b4-6823-47ef-a8a6-52c94e53ef4c\") " pod="openshift-console/console-f9d7485db-4bxkk" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324635 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71145266-38d8-495b-83c2-cff3c68e705d-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2ftds\" (UID: \"71145266-38d8-495b-83c2-cff3c68e705d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ftds" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324660 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/710fc779-629b-4d37-a820-e08f5289677f-audit-dir\") pod \"apiserver-76f77b778f-tgwt9\" (UID: \"710fc779-629b-4d37-a820-e08f5289677f\") " pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324687 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xczsj\" (UniqueName: \"kubernetes.io/projected/dcf21347-4c2f-4d66-9137-c24ee63a31de-kube-api-access-xczsj\") pod \"csi-hostpathplugin-7jgjj\" (UID: \"dcf21347-4c2f-4d66-9137-c24ee63a31de\") " pod="hostpath-provisioner/csi-hostpathplugin-7jgjj" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324734 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9feb8b3f-6359-41f1-8648-c827ce9c5ec4-metrics-tls\") pod \"dns-default-5795c\" (UID: \"9feb8b3f-6359-41f1-8648-c827ce9c5ec4\") " pod="openshift-dns/dns-default-5795c" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324758 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fa6c59b4-6823-47ef-a8a6-52c94e53ef4c-service-ca\") pod \"console-f9d7485db-4bxkk\" (UID: \"fa6c59b4-6823-47ef-a8a6-52c94e53ef4c\") " pod="openshift-console/console-f9d7485db-4bxkk" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324791 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/31b0687b-5c42-4259-b96d-284c1aa487b4-client-ca\") pod \"controller-manager-879f6c89f-27bbf\" (UID: \"31b0687b-5c42-4259-b96d-284c1aa487b4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-27bbf" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324814 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324837 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2jpm\" (UniqueName: \"kubernetes.io/projected/5c0e215f-d940-415a-90d3-f5faf3e537de-kube-api-access-j2jpm\") pod \"olm-operator-6b444d44fb-j99bk\" (UID: \"5c0e215f-d940-415a-90d3-f5faf3e537de\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j99bk" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324902 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324929 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/3cd5b209-b64c-4ce1-85e2-e00e568b4bbe-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-bw75d\" (UID: \"3cd5b209-b64c-4ce1-85e2-e00e568b4bbe\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bw75d" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.324960 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwt8v\" (UniqueName: \"kubernetes.io/projected/0f95dce4-6d68-44ac-9dd3-215664bb1d75-kube-api-access-lwt8v\") pod \"marketplace-operator-79b997595-xrhcj\" (UID: \"0f95dce4-6d68-44ac-9dd3-215664bb1d75\") " pod="openshift-marketplace/marketplace-operator-79b997595-xrhcj" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325012 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/21fb3328-3882-4644-9a94-b7949f48db54-images\") pod \"machine-config-operator-74547568cd-6frkl\" (UID: \"21fb3328-3882-4644-9a94-b7949f48db54\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6frkl" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325037 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5c0e215f-d940-415a-90d3-f5faf3e537de-srv-cert\") pod \"olm-operator-6b444d44fb-j99bk\" (UID: \"5c0e215f-d940-415a-90d3-f5faf3e537de\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j99bk" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325065 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325081 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325108 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-272j4\" (UniqueName: \"kubernetes.io/projected/31b0687b-5c42-4259-b96d-284c1aa487b4-kube-api-access-272j4\") pod \"controller-manager-879f6c89f-27bbf\" (UID: \"31b0687b-5c42-4259-b96d-284c1aa487b4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-27bbf" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325124 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/86b5fc8f-1b76-43ad-b283-0d46697be7aa-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hv6vt\" (UID: \"86b5fc8f-1b76-43ad-b283-0d46697be7aa\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hv6vt" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325141 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/dcf21347-4c2f-4d66-9137-c24ee63a31de-registration-dir\") pod \"csi-hostpathplugin-7jgjj\" (UID: \"dcf21347-4c2f-4d66-9137-c24ee63a31de\") " pod="hostpath-provisioner/csi-hostpathplugin-7jgjj" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325182 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/710fc779-629b-4d37-a820-e08f5289677f-audit\") pod \"apiserver-76f77b778f-tgwt9\" (UID: \"710fc779-629b-4d37-a820-e08f5289677f\") " pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325198 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/8db0982f-0e1d-493c-8b32-64190372fcb9-signing-key\") pod \"service-ca-9c57cc56f-ccl77\" (UID: \"8db0982f-0e1d-493c-8b32-64190372fcb9\") " pod="openshift-service-ca/service-ca-9c57cc56f-ccl77" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325214 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/6225d3ae-2a4e-479c-aa70-6fdbd2ee488b-etcd-service-ca\") pod \"etcd-operator-b45778765-fhpqt\" (UID: \"6225d3ae-2a4e-479c-aa70-6fdbd2ee488b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fhpqt" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325269 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6225d3ae-2a4e-479c-aa70-6fdbd2ee488b-etcd-client\") pod \"etcd-operator-b45778765-fhpqt\" (UID: \"6225d3ae-2a4e-479c-aa70-6fdbd2ee488b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fhpqt" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325290 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31b0687b-5c42-4259-b96d-284c1aa487b4-serving-cert\") pod \"controller-manager-879f6c89f-27bbf\" (UID: \"31b0687b-5c42-4259-b96d-284c1aa487b4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-27bbf" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325306 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xf69f\" (UniqueName: \"kubernetes.io/projected/8fa6844e-65a5-401e-b329-2a49dc5debe4-kube-api-access-xf69f\") pod \"ingress-canary-znxvs\" (UID: \"8fa6844e-65a5-401e-b329-2a49dc5debe4\") " pod="openshift-ingress-canary/ingress-canary-znxvs" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325322 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b60354fe-e785-4e88-af1c-0d846d9ff59a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bjbsw\" (UID: \"b60354fe-e785-4e88-af1c-0d846d9ff59a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bjbsw" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325348 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txk8c\" (UniqueName: \"kubernetes.io/projected/6225d3ae-2a4e-479c-aa70-6fdbd2ee488b-kube-api-access-txk8c\") pod \"etcd-operator-b45778765-fhpqt\" (UID: \"6225d3ae-2a4e-479c-aa70-6fdbd2ee488b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fhpqt" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325373 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/351aa26e-939a-4b69-a7ee-55be7111ee6e-ca-trust-extracted\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325403 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/710fc779-629b-4d37-a820-e08f5289677f-config\") pod \"apiserver-76f77b778f-tgwt9\" (UID: \"710fc779-629b-4d37-a820-e08f5289677f\") " pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325418 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/64e32b45-d067-4151-980d-36ea8f3116e9-srv-cert\") pod \"catalog-operator-68c6474976-66zwg\" (UID: \"64e32b45-d067-4151-980d-36ea8f3116e9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-66zwg" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325443 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37bf7cb8-e127-458a-ab07-efb2c111ce02-metrics-tls\") pod \"ingress-operator-5b745b69d9-dlghr\" (UID: \"37bf7cb8-e127-458a-ab07-efb2c111ce02\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dlghr" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325459 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325484 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r9vl\" (UniqueName: \"kubernetes.io/projected/9feb8b3f-6359-41f1-8648-c827ce9c5ec4-kube-api-access-4r9vl\") pod \"dns-default-5795c\" (UID: \"9feb8b3f-6359-41f1-8648-c827ce9c5ec4\") " pod="openshift-dns/dns-default-5795c" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325499 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7l5m5\" (UniqueName: \"kubernetes.io/projected/64e32b45-d067-4151-980d-36ea8f3116e9-kube-api-access-7l5m5\") pod \"catalog-operator-68c6474976-66zwg\" (UID: \"64e32b45-d067-4151-980d-36ea8f3116e9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-66zwg" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325516 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f88rx\" (UniqueName: \"kubernetes.io/projected/351aa26e-939a-4b69-a7ee-55be7111ee6e-kube-api-access-f88rx\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325533 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23b10824-dc70-459f-b037-6f8f00ae6a04-serving-cert\") pod \"openshift-config-operator-7777fb866f-mcjhx\" (UID: \"23b10824-dc70-459f-b037-6f8f00ae6a04\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mcjhx" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325548 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b82p9\" (UniqueName: \"kubernetes.io/projected/8db0982f-0e1d-493c-8b32-64190372fcb9-kube-api-access-b82p9\") pod \"service-ca-9c57cc56f-ccl77\" (UID: \"8db0982f-0e1d-493c-8b32-64190372fcb9\") " pod="openshift-service-ca/service-ca-9c57cc56f-ccl77" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325565 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99b5p\" (UniqueName: \"kubernetes.io/projected/013d2ce9-60ac-4abf-945d-3340ffa55b72-kube-api-access-99b5p\") pod \"machine-config-server-s95mb\" (UID: \"013d2ce9-60ac-4abf-945d-3340ffa55b72\") " pod="openshift-machine-config-operator/machine-config-server-s95mb" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325596 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/351aa26e-939a-4b69-a7ee-55be7111ee6e-installation-pull-secrets\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325612 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/351aa26e-939a-4b69-a7ee-55be7111ee6e-registry-tls\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325627 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/351aa26e-939a-4b69-a7ee-55be7111ee6e-registry-certificates\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325644 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bjgv\" (UniqueName: \"kubernetes.io/projected/036e7602-7608-4033-b18d-d859a7eddebd-kube-api-access-4bjgv\") pod \"downloads-7954f5f757-6v4cn\" (UID: \"036e7602-7608-4033-b18d-d859a7eddebd\") " pod="openshift-console/downloads-7954f5f757-6v4cn" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325659 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9feb8b3f-6359-41f1-8648-c827ce9c5ec4-config-volume\") pod \"dns-default-5795c\" (UID: \"9feb8b3f-6359-41f1-8648-c827ce9c5ec4\") " pod="openshift-dns/dns-default-5795c" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325676 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6225d3ae-2a4e-479c-aa70-6fdbd2ee488b-config\") pod \"etcd-operator-b45778765-fhpqt\" (UID: \"6225d3ae-2a4e-479c-aa70-6fdbd2ee488b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fhpqt" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325695 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/76cb1e0e-4321-4450-9fcb-b505a7a1d969-audit-policies\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325713 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/8c1a0565-e55b-410e-9b63-12ababdb69eb-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-jpg48\" (UID: \"8c1a0565-e55b-410e-9b63-12ababdb69eb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jpg48" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325730 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/6225d3ae-2a4e-479c-aa70-6fdbd2ee488b-etcd-ca\") pod \"etcd-operator-b45778765-fhpqt\" (UID: \"6225d3ae-2a4e-479c-aa70-6fdbd2ee488b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fhpqt" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325747 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4122c286-47e3-460d-bd25-323012847f12-config\") pod \"kube-controller-manager-operator-78b949d7b-2bncs\" (UID: \"4122c286-47e3-460d-bd25-323012847f12\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bncs" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325761 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/dcf21347-4c2f-4d66-9137-c24ee63a31de-socket-dir\") pod \"csi-hostpathplugin-7jgjj\" (UID: \"dcf21347-4c2f-4d66-9137-c24ee63a31de\") " pod="hostpath-provisioner/csi-hostpathplugin-7jgjj" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325781 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/710fc779-629b-4d37-a820-e08f5289677f-node-pullsecrets\") pod \"apiserver-76f77b778f-tgwt9\" (UID: \"710fc779-629b-4d37-a820-e08f5289677f\") " pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325800 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/fa6c59b4-6823-47ef-a8a6-52c94e53ef4c-console-oauth-config\") pod \"console-f9d7485db-4bxkk\" (UID: \"fa6c59b4-6823-47ef-a8a6-52c94e53ef4c\") " pod="openshift-console/console-f9d7485db-4bxkk" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325816 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tq74b\" (UniqueName: \"kubernetes.io/projected/dd10c491-df5d-43a1-961d-613e2a90be23-kube-api-access-tq74b\") pod \"collect-profiles-29523765-frmvz\" (UID: \"dd10c491-df5d-43a1-961d-613e2a90be23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523765-frmvz" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325843 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7b9d6a43-e957-4d02-9928-020f9f17f7a9-apiservice-cert\") pod \"packageserver-d55dfcdfc-l8fcf\" (UID: \"7b9d6a43-e957-4d02-9928-020f9f17f7a9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l8fcf" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325857 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7b9d6a43-e957-4d02-9928-020f9f17f7a9-webhook-cert\") pod \"packageserver-d55dfcdfc-l8fcf\" (UID: \"7b9d6a43-e957-4d02-9928-020f9f17f7a9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l8fcf" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325871 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ck9fl\" (UniqueName: \"kubernetes.io/projected/8c1a0565-e55b-410e-9b63-12ababdb69eb-kube-api-access-ck9fl\") pod \"control-plane-machine-set-operator-78cbb6b69f-jpg48\" (UID: \"8c1a0565-e55b-410e-9b63-12ababdb69eb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jpg48" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325902 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325917 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/710fc779-629b-4d37-a820-e08f5289677f-image-import-ca\") pod \"apiserver-76f77b778f-tgwt9\" (UID: \"710fc779-629b-4d37-a820-e08f5289677f\") " pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325934 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlg8k\" (UniqueName: \"kubernetes.io/projected/fa6c59b4-6823-47ef-a8a6-52c94e53ef4c-kube-api-access-wlg8k\") pod \"console-f9d7485db-4bxkk\" (UID: \"fa6c59b4-6823-47ef-a8a6-52c94e53ef4c\") " pod="openshift-console/console-f9d7485db-4bxkk" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325951 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/17344065-a988-48bb-aff9-d51ffd76f54b-metrics-tls\") pod \"dns-operator-744455d44c-rmc7c\" (UID: \"17344065-a988-48bb-aff9-d51ffd76f54b\") " pod="openshift-dns-operator/dns-operator-744455d44c-rmc7c" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325975 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqvw5\" (UniqueName: \"kubernetes.io/projected/7b9d6a43-e957-4d02-9928-020f9f17f7a9-kube-api-access-lqvw5\") pod \"packageserver-d55dfcdfc-l8fcf\" (UID: \"7b9d6a43-e957-4d02-9928-020f9f17f7a9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l8fcf" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.325993 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.326009 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b60354fe-e785-4e88-af1c-0d846d9ff59a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bjbsw\" (UID: \"b60354fe-e785-4e88-af1c-0d846d9ff59a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bjbsw" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.326025 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.326042 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqzfn\" (UniqueName: \"kubernetes.io/projected/17344065-a988-48bb-aff9-d51ffd76f54b-kube-api-access-cqzfn\") pod \"dns-operator-744455d44c-rmc7c\" (UID: \"17344065-a988-48bb-aff9-d51ffd76f54b\") " pod="openshift-dns-operator/dns-operator-744455d44c-rmc7c" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.326069 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/710fc779-629b-4d37-a820-e08f5289677f-trusted-ca-bundle\") pod \"apiserver-76f77b778f-tgwt9\" (UID: \"710fc779-629b-4d37-a820-e08f5289677f\") " pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.326317 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f8abcf75-9577-4dfe-a252-40aff339ffce-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-gqb6w\" (UID: \"f8abcf75-9577-4dfe-a252-40aff339ffce\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gqb6w" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.326383 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/351aa26e-939a-4b69-a7ee-55be7111ee6e-ca-trust-extracted\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.327153 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/710fc779-629b-4d37-a820-e08f5289677f-config\") pod \"apiserver-76f77b778f-tgwt9\" (UID: \"710fc779-629b-4d37-a820-e08f5289677f\") " pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.326078 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/710fc779-629b-4d37-a820-e08f5289677f-serving-cert\") pod \"apiserver-76f77b778f-tgwt9\" (UID: \"710fc779-629b-4d37-a820-e08f5289677f\") " pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.327225 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/710fc779-629b-4d37-a820-e08f5289677f-encryption-config\") pod \"apiserver-76f77b778f-tgwt9\" (UID: \"710fc779-629b-4d37-a820-e08f5289677f\") " pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.327578 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/23b10824-dc70-459f-b037-6f8f00ae6a04-available-featuregates\") pod \"openshift-config-operator-7777fb866f-mcjhx\" (UID: \"23b10824-dc70-459f-b037-6f8f00ae6a04\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mcjhx" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.328086 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/21fb3328-3882-4644-9a94-b7949f48db54-images\") pod \"machine-config-operator-74547568cd-6frkl\" (UID: \"21fb3328-3882-4644-9a94-b7949f48db54\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6frkl" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.328295 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/710fc779-629b-4d37-a820-e08f5289677f-audit\") pod \"apiserver-76f77b778f-tgwt9\" (UID: \"710fc779-629b-4d37-a820-e08f5289677f\") " pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: E0218 14:57:24.328534 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:24.828518649 +0000 UTC m=+143.757852962 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.328793 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fa6c59b4-6823-47ef-a8a6-52c94e53ef4c-service-ca\") pod \"console-f9d7485db-4bxkk\" (UID: \"fa6c59b4-6823-47ef-a8a6-52c94e53ef4c\") " pod="openshift-console/console-f9d7485db-4bxkk" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.329608 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.329641 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/3cd5b209-b64c-4ce1-85e2-e00e568b4bbe-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-bw75d\" (UID: \"3cd5b209-b64c-4ce1-85e2-e00e568b4bbe\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bw75d" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.329994 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/918ef2a1-8828-4b17-b30a-de957f009104-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-z7xdz\" (UID: \"918ef2a1-8828-4b17-b30a-de957f009104\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-z7xdz" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.330383 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.330766 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31b0687b-5c42-4259-b96d-284c1aa487b4-serving-cert\") pod \"controller-manager-879f6c89f-27bbf\" (UID: \"31b0687b-5c42-4259-b96d-284c1aa487b4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-27bbf" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.331440 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/710fc779-629b-4d37-a820-e08f5289677f-serving-cert\") pod \"apiserver-76f77b778f-tgwt9\" (UID: \"710fc779-629b-4d37-a820-e08f5289677f\") " pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.332458 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/710fc779-629b-4d37-a820-e08f5289677f-encryption-config\") pod \"apiserver-76f77b778f-tgwt9\" (UID: \"710fc779-629b-4d37-a820-e08f5289677f\") " pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.332501 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.332886 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/710fc779-629b-4d37-a820-e08f5289677f-node-pullsecrets\") pod \"apiserver-76f77b778f-tgwt9\" (UID: \"710fc779-629b-4d37-a820-e08f5289677f\") " pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.333440 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31b0687b-5c42-4259-b96d-284c1aa487b4-config\") pod \"controller-manager-879f6c89f-27bbf\" (UID: \"31b0687b-5c42-4259-b96d-284c1aa487b4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-27bbf" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.334008 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/710fc779-629b-4d37-a820-e08f5289677f-audit-dir\") pod \"apiserver-76f77b778f-tgwt9\" (UID: \"710fc779-629b-4d37-a820-e08f5289677f\") " pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.334449 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.334746 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/710fc779-629b-4d37-a820-e08f5289677f-image-import-ca\") pod \"apiserver-76f77b778f-tgwt9\" (UID: \"710fc779-629b-4d37-a820-e08f5289677f\") " pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.335411 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/351aa26e-939a-4b69-a7ee-55be7111ee6e-registry-certificates\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.336046 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/76cb1e0e-4321-4450-9fcb-b505a7a1d969-audit-policies\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.336331 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-8cq46" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.337440 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/710fc779-629b-4d37-a820-e08f5289677f-etcd-client\") pod \"apiserver-76f77b778f-tgwt9\" (UID: \"710fc779-629b-4d37-a820-e08f5289677f\") " pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.337842 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/31b0687b-5c42-4259-b96d-284c1aa487b4-client-ca\") pod \"controller-manager-879f6c89f-27bbf\" (UID: \"31b0687b-5c42-4259-b96d-284c1aa487b4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-27bbf" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.340463 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.342188 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.342575 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/351aa26e-939a-4b69-a7ee-55be7111ee6e-registry-tls\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.345769 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.346027 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/351aa26e-939a-4b69-a7ee-55be7111ee6e-installation-pull-secrets\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.346336 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37bf7cb8-e127-458a-ab07-efb2c111ce02-metrics-tls\") pod \"ingress-operator-5b745b69d9-dlghr\" (UID: \"37bf7cb8-e127-458a-ab07-efb2c111ce02\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dlghr" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.347610 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.347702 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/fa6c59b4-6823-47ef-a8a6-52c94e53ef4c-console-oauth-config\") pod \"console-f9d7485db-4bxkk\" (UID: \"fa6c59b4-6823-47ef-a8a6-52c94e53ef4c\") " pod="openshift-console/console-f9d7485db-4bxkk" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.347802 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/17344065-a988-48bb-aff9-d51ffd76f54b-metrics-tls\") pod \"dns-operator-744455d44c-rmc7c\" (UID: \"17344065-a988-48bb-aff9-d51ffd76f54b\") " pod="openshift-dns-operator/dns-operator-744455d44c-rmc7c" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.347816 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71145266-38d8-495b-83c2-cff3c68e705d-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2ftds\" (UID: \"71145266-38d8-495b-83c2-cff3c68e705d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ftds" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.349593 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfwkt\" (UniqueName: \"kubernetes.io/projected/23b10824-dc70-459f-b037-6f8f00ae6a04-kube-api-access-cfwkt\") pod \"openshift-config-operator-7777fb866f-mcjhx\" (UID: \"23b10824-dc70-459f-b037-6f8f00ae6a04\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mcjhx" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.355251 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-966wr" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.359159 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23b10824-dc70-459f-b037-6f8f00ae6a04-serving-cert\") pod \"openshift-config-operator-7777fb866f-mcjhx\" (UID: \"23b10824-dc70-459f-b037-6f8f00ae6a04\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mcjhx" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.374465 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dl5w6\" (UniqueName: \"kubernetes.io/projected/37bf7cb8-e127-458a-ab07-efb2c111ce02-kube-api-access-dl5w6\") pod \"ingress-operator-5b745b69d9-dlghr\" (UID: \"37bf7cb8-e127-458a-ab07-efb2c111ce02\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dlghr" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.388824 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/351aa26e-939a-4b69-a7ee-55be7111ee6e-bound-sa-token\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:24 crc kubenswrapper[4896]: W0218 14:57:24.408487 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a51266e_326e_4b5c_a25a_fb61bf1f3757.slice/crio-9071936d0faf6e8ddcc4832ae2d941a12b602f63bb32ba26985268b3ae6ceae8 WatchSource:0}: Error finding container 9071936d0faf6e8ddcc4832ae2d941a12b602f63bb32ba26985268b3ae6ceae8: Status 404 returned error can't find the container with id 9071936d0faf6e8ddcc4832ae2d941a12b602f63bb32ba26985268b3ae6ceae8 Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.411848 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/37bf7cb8-e127-458a-ab07-efb2c111ce02-bound-sa-token\") pod \"ingress-operator-5b745b69d9-dlghr\" (UID: \"37bf7cb8-e127-458a-ab07-efb2c111ce02\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dlghr" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.428526 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.428688 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/8db0982f-0e1d-493c-8b32-64190372fcb9-signing-key\") pod \"service-ca-9c57cc56f-ccl77\" (UID: \"8db0982f-0e1d-493c-8b32-64190372fcb9\") " pod="openshift-service-ca/service-ca-9c57cc56f-ccl77" Feb 18 14:57:24 crc kubenswrapper[4896]: E0218 14:57:24.428717 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:24.928684107 +0000 UTC m=+143.858018320 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.428762 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/6225d3ae-2a4e-479c-aa70-6fdbd2ee488b-etcd-service-ca\") pod \"etcd-operator-b45778765-fhpqt\" (UID: \"6225d3ae-2a4e-479c-aa70-6fdbd2ee488b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fhpqt" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.428811 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6225d3ae-2a4e-479c-aa70-6fdbd2ee488b-etcd-client\") pod \"etcd-operator-b45778765-fhpqt\" (UID: \"6225d3ae-2a4e-479c-aa70-6fdbd2ee488b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fhpqt" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.428837 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xf69f\" (UniqueName: \"kubernetes.io/projected/8fa6844e-65a5-401e-b329-2a49dc5debe4-kube-api-access-xf69f\") pod \"ingress-canary-znxvs\" (UID: \"8fa6844e-65a5-401e-b329-2a49dc5debe4\") " pod="openshift-ingress-canary/ingress-canary-znxvs" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.428860 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b60354fe-e785-4e88-af1c-0d846d9ff59a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bjbsw\" (UID: \"b60354fe-e785-4e88-af1c-0d846d9ff59a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bjbsw" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.428886 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txk8c\" (UniqueName: \"kubernetes.io/projected/6225d3ae-2a4e-479c-aa70-6fdbd2ee488b-kube-api-access-txk8c\") pod \"etcd-operator-b45778765-fhpqt\" (UID: \"6225d3ae-2a4e-479c-aa70-6fdbd2ee488b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fhpqt" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.429505 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/64e32b45-d067-4151-980d-36ea8f3116e9-srv-cert\") pod \"catalog-operator-68c6474976-66zwg\" (UID: \"64e32b45-d067-4151-980d-36ea8f3116e9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-66zwg" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.429574 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4r9vl\" (UniqueName: \"kubernetes.io/projected/9feb8b3f-6359-41f1-8648-c827ce9c5ec4-kube-api-access-4r9vl\") pod \"dns-default-5795c\" (UID: \"9feb8b3f-6359-41f1-8648-c827ce9c5ec4\") " pod="openshift-dns/dns-default-5795c" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.429603 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7l5m5\" (UniqueName: \"kubernetes.io/projected/64e32b45-d067-4151-980d-36ea8f3116e9-kube-api-access-7l5m5\") pod \"catalog-operator-68c6474976-66zwg\" (UID: \"64e32b45-d067-4151-980d-36ea8f3116e9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-66zwg" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.429665 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b82p9\" (UniqueName: \"kubernetes.io/projected/8db0982f-0e1d-493c-8b32-64190372fcb9-kube-api-access-b82p9\") pod \"service-ca-9c57cc56f-ccl77\" (UID: \"8db0982f-0e1d-493c-8b32-64190372fcb9\") " pod="openshift-service-ca/service-ca-9c57cc56f-ccl77" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.429691 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99b5p\" (UniqueName: \"kubernetes.io/projected/013d2ce9-60ac-4abf-945d-3340ffa55b72-kube-api-access-99b5p\") pod \"machine-config-server-s95mb\" (UID: \"013d2ce9-60ac-4abf-945d-3340ffa55b72\") " pod="openshift-machine-config-operator/machine-config-server-s95mb" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.429720 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/6225d3ae-2a4e-479c-aa70-6fdbd2ee488b-etcd-service-ca\") pod \"etcd-operator-b45778765-fhpqt\" (UID: \"6225d3ae-2a4e-479c-aa70-6fdbd2ee488b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fhpqt" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.429762 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9feb8b3f-6359-41f1-8648-c827ce9c5ec4-config-volume\") pod \"dns-default-5795c\" (UID: \"9feb8b3f-6359-41f1-8648-c827ce9c5ec4\") " pod="openshift-dns/dns-default-5795c" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.429789 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6225d3ae-2a4e-479c-aa70-6fdbd2ee488b-config\") pod \"etcd-operator-b45778765-fhpqt\" (UID: \"6225d3ae-2a4e-479c-aa70-6fdbd2ee488b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fhpqt" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.429846 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/8c1a0565-e55b-410e-9b63-12ababdb69eb-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-jpg48\" (UID: \"8c1a0565-e55b-410e-9b63-12ababdb69eb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jpg48" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.429869 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/6225d3ae-2a4e-479c-aa70-6fdbd2ee488b-etcd-ca\") pod \"etcd-operator-b45778765-fhpqt\" (UID: \"6225d3ae-2a4e-479c-aa70-6fdbd2ee488b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fhpqt" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.429923 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4122c286-47e3-460d-bd25-323012847f12-config\") pod \"kube-controller-manager-operator-78b949d7b-2bncs\" (UID: \"4122c286-47e3-460d-bd25-323012847f12\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bncs" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.429948 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/dcf21347-4c2f-4d66-9137-c24ee63a31de-socket-dir\") pod \"csi-hostpathplugin-7jgjj\" (UID: \"dcf21347-4c2f-4d66-9137-c24ee63a31de\") " pod="hostpath-provisioner/csi-hostpathplugin-7jgjj" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.430001 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tq74b\" (UniqueName: \"kubernetes.io/projected/dd10c491-df5d-43a1-961d-613e2a90be23-kube-api-access-tq74b\") pod \"collect-profiles-29523765-frmvz\" (UID: \"dd10c491-df5d-43a1-961d-613e2a90be23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523765-frmvz" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.430028 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7b9d6a43-e957-4d02-9928-020f9f17f7a9-apiservice-cert\") pod \"packageserver-d55dfcdfc-l8fcf\" (UID: \"7b9d6a43-e957-4d02-9928-020f9f17f7a9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l8fcf" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.430080 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7b9d6a43-e957-4d02-9928-020f9f17f7a9-webhook-cert\") pod \"packageserver-d55dfcdfc-l8fcf\" (UID: \"7b9d6a43-e957-4d02-9928-020f9f17f7a9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l8fcf" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.430108 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ck9fl\" (UniqueName: \"kubernetes.io/projected/8c1a0565-e55b-410e-9b63-12ababdb69eb-kube-api-access-ck9fl\") pod \"control-plane-machine-set-operator-78cbb6b69f-jpg48\" (UID: \"8c1a0565-e55b-410e-9b63-12ababdb69eb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jpg48" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.430180 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqvw5\" (UniqueName: \"kubernetes.io/projected/7b9d6a43-e957-4d02-9928-020f9f17f7a9-kube-api-access-lqvw5\") pod \"packageserver-d55dfcdfc-l8fcf\" (UID: \"7b9d6a43-e957-4d02-9928-020f9f17f7a9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l8fcf" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.430209 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b60354fe-e785-4e88-af1c-0d846d9ff59a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bjbsw\" (UID: \"b60354fe-e785-4e88-af1c-0d846d9ff59a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bjbsw" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.430494 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6225d3ae-2a4e-479c-aa70-6fdbd2ee488b-config\") pod \"etcd-operator-b45778765-fhpqt\" (UID: \"6225d3ae-2a4e-479c-aa70-6fdbd2ee488b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fhpqt" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.430524 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/dcf21347-4c2f-4d66-9137-c24ee63a31de-mountpoint-dir\") pod \"csi-hostpathplugin-7jgjj\" (UID: \"dcf21347-4c2f-4d66-9137-c24ee63a31de\") " pod="hostpath-provisioner/csi-hostpathplugin-7jgjj" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.430585 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b60354fe-e785-4e88-af1c-0d846d9ff59a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bjbsw\" (UID: \"b60354fe-e785-4e88-af1c-0d846d9ff59a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bjbsw" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.430643 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0f95dce4-6d68-44ac-9dd3-215664bb1d75-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xrhcj\" (UID: \"0f95dce4-6d68-44ac-9dd3-215664bb1d75\") " pod="openshift-marketplace/marketplace-operator-79b997595-xrhcj" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.430668 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dd10c491-df5d-43a1-961d-613e2a90be23-config-volume\") pod \"collect-profiles-29523765-frmvz\" (UID: \"dd10c491-df5d-43a1-961d-613e2a90be23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523765-frmvz" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.430689 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dd10c491-df5d-43a1-961d-613e2a90be23-secret-volume\") pod \"collect-profiles-29523765-frmvz\" (UID: \"dd10c491-df5d-43a1-961d-613e2a90be23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523765-frmvz" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.430745 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2573759f-2db1-452a-b310-f2a170099ec6-serving-cert\") pod \"service-ca-operator-777779d784-rx8f2\" (UID: \"2573759f-2db1-452a-b310-f2a170099ec6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-rx8f2" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.430805 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0f95dce4-6d68-44ac-9dd3-215664bb1d75-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xrhcj\" (UID: \"0f95dce4-6d68-44ac-9dd3-215664bb1d75\") " pod="openshift-marketplace/marketplace-operator-79b997595-xrhcj" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.430830 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/013d2ce9-60ac-4abf-945d-3340ffa55b72-node-bootstrap-token\") pod \"machine-config-server-s95mb\" (UID: \"013d2ce9-60ac-4abf-945d-3340ffa55b72\") " pod="openshift-machine-config-operator/machine-config-server-s95mb" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.430853 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/dcf21347-4c2f-4d66-9137-c24ee63a31de-csi-data-dir\") pod \"csi-hostpathplugin-7jgjj\" (UID: \"dcf21347-4c2f-4d66-9137-c24ee63a31de\") " pod="hostpath-provisioner/csi-hostpathplugin-7jgjj" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.430875 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/6225d3ae-2a4e-479c-aa70-6fdbd2ee488b-etcd-ca\") pod \"etcd-operator-b45778765-fhpqt\" (UID: \"6225d3ae-2a4e-479c-aa70-6fdbd2ee488b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fhpqt" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.430880 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/8db0982f-0e1d-493c-8b32-64190372fcb9-signing-cabundle\") pod \"service-ca-9c57cc56f-ccl77\" (UID: \"8db0982f-0e1d-493c-8b32-64190372fcb9\") " pod="openshift-service-ca/service-ca-9c57cc56f-ccl77" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.430926 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/7b9d6a43-e957-4d02-9928-020f9f17f7a9-tmpfs\") pod \"packageserver-d55dfcdfc-l8fcf\" (UID: \"7b9d6a43-e957-4d02-9928-020f9f17f7a9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l8fcf" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.430966 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4122c286-47e3-460d-bd25-323012847f12-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-2bncs\" (UID: \"4122c286-47e3-460d-bd25-323012847f12\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bncs" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.431005 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5c0e215f-d940-415a-90d3-f5faf3e537de-profile-collector-cert\") pod \"olm-operator-6b444d44fb-j99bk\" (UID: \"5c0e215f-d940-415a-90d3-f5faf3e537de\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j99bk" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.431027 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6225d3ae-2a4e-479c-aa70-6fdbd2ee488b-serving-cert\") pod \"etcd-operator-b45778765-fhpqt\" (UID: \"6225d3ae-2a4e-479c-aa70-6fdbd2ee488b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fhpqt" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.431052 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jnkn\" (UniqueName: \"kubernetes.io/projected/b0388f1f-8b9c-46e4-a7f0-1fad187d6d65-kube-api-access-5jnkn\") pod \"machine-config-controller-84d6567774-v59s8\" (UID: \"b0388f1f-8b9c-46e4-a7f0-1fad187d6d65\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v59s8" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.431077 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgmr5\" (UniqueName: \"kubernetes.io/projected/0e9fc2ab-6d3d-441d-8591-3d64ce77c62d-kube-api-access-pgmr5\") pod \"migrator-59844c95c7-pclrg\" (UID: \"0e9fc2ab-6d3d-441d-8591-3d64ce77c62d\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pclrg" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.431116 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4122c286-47e3-460d-bd25-323012847f12-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-2bncs\" (UID: \"4122c286-47e3-460d-bd25-323012847f12\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bncs" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.431143 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8fa6844e-65a5-401e-b329-2a49dc5debe4-cert\") pod \"ingress-canary-znxvs\" (UID: \"8fa6844e-65a5-401e-b329-2a49dc5debe4\") " pod="openshift-ingress-canary/ingress-canary-znxvs" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.431164 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/013d2ce9-60ac-4abf-945d-3340ffa55b72-certs\") pod \"machine-config-server-s95mb\" (UID: \"013d2ce9-60ac-4abf-945d-3340ffa55b72\") " pod="openshift-machine-config-operator/machine-config-server-s95mb" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.431184 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/dcf21347-4c2f-4d66-9137-c24ee63a31de-plugins-dir\") pod \"csi-hostpathplugin-7jgjj\" (UID: \"dcf21347-4c2f-4d66-9137-c24ee63a31de\") " pod="hostpath-provisioner/csi-hostpathplugin-7jgjj" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.431225 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b0388f1f-8b9c-46e4-a7f0-1fad187d6d65-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-v59s8\" (UID: \"b0388f1f-8b9c-46e4-a7f0-1fad187d6d65\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v59s8" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.431309 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b0388f1f-8b9c-46e4-a7f0-1fad187d6d65-proxy-tls\") pod \"machine-config-controller-84d6567774-v59s8\" (UID: \"b0388f1f-8b9c-46e4-a7f0-1fad187d6d65\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v59s8" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.431363 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q5kn\" (UniqueName: \"kubernetes.io/projected/2573759f-2db1-452a-b310-f2a170099ec6-kube-api-access-8q5kn\") pod \"service-ca-operator-777779d784-rx8f2\" (UID: \"2573759f-2db1-452a-b310-f2a170099ec6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-rx8f2" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.431398 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/64e32b45-d067-4151-980d-36ea8f3116e9-profile-collector-cert\") pod \"catalog-operator-68c6474976-66zwg\" (UID: \"64e32b45-d067-4151-980d-36ea8f3116e9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-66zwg" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.431420 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2573759f-2db1-452a-b310-f2a170099ec6-config\") pod \"service-ca-operator-777779d784-rx8f2\" (UID: \"2573759f-2db1-452a-b310-f2a170099ec6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-rx8f2" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.431442 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rbcf\" (UniqueName: \"kubernetes.io/projected/86b5fc8f-1b76-43ad-b283-0d46697be7aa-kube-api-access-5rbcf\") pod \"package-server-manager-789f6589d5-hv6vt\" (UID: \"86b5fc8f-1b76-43ad-b283-0d46697be7aa\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hv6vt" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.431467 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xczsj\" (UniqueName: \"kubernetes.io/projected/dcf21347-4c2f-4d66-9137-c24ee63a31de-kube-api-access-xczsj\") pod \"csi-hostpathplugin-7jgjj\" (UID: \"dcf21347-4c2f-4d66-9137-c24ee63a31de\") " pod="hostpath-provisioner/csi-hostpathplugin-7jgjj" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.431488 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9feb8b3f-6359-41f1-8648-c827ce9c5ec4-metrics-tls\") pod \"dns-default-5795c\" (UID: \"9feb8b3f-6359-41f1-8648-c827ce9c5ec4\") " pod="openshift-dns/dns-default-5795c" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.431509 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2jpm\" (UniqueName: \"kubernetes.io/projected/5c0e215f-d940-415a-90d3-f5faf3e537de-kube-api-access-j2jpm\") pod \"olm-operator-6b444d44fb-j99bk\" (UID: \"5c0e215f-d940-415a-90d3-f5faf3e537de\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j99bk" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.431517 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9feb8b3f-6359-41f1-8648-c827ce9c5ec4-config-volume\") pod \"dns-default-5795c\" (UID: \"9feb8b3f-6359-41f1-8648-c827ce9c5ec4\") " pod="openshift-dns/dns-default-5795c" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.431539 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.431576 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwt8v\" (UniqueName: \"kubernetes.io/projected/0f95dce4-6d68-44ac-9dd3-215664bb1d75-kube-api-access-lwt8v\") pod \"marketplace-operator-79b997595-xrhcj\" (UID: \"0f95dce4-6d68-44ac-9dd3-215664bb1d75\") " pod="openshift-marketplace/marketplace-operator-79b997595-xrhcj" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.431602 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5c0e215f-d940-415a-90d3-f5faf3e537de-srv-cert\") pod \"olm-operator-6b444d44fb-j99bk\" (UID: \"5c0e215f-d940-415a-90d3-f5faf3e537de\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j99bk" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.431672 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/86b5fc8f-1b76-43ad-b283-0d46697be7aa-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hv6vt\" (UID: \"86b5fc8f-1b76-43ad-b283-0d46697be7aa\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hv6vt" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.431703 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/dcf21347-4c2f-4d66-9137-c24ee63a31de-registration-dir\") pod \"csi-hostpathplugin-7jgjj\" (UID: \"dcf21347-4c2f-4d66-9137-c24ee63a31de\") " pod="hostpath-provisioner/csi-hostpathplugin-7jgjj" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.431814 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/dcf21347-4c2f-4d66-9137-c24ee63a31de-socket-dir\") pod \"csi-hostpathplugin-7jgjj\" (UID: \"dcf21347-4c2f-4d66-9137-c24ee63a31de\") " pod="hostpath-provisioner/csi-hostpathplugin-7jgjj" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.431840 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/dcf21347-4c2f-4d66-9137-c24ee63a31de-registration-dir\") pod \"csi-hostpathplugin-7jgjj\" (UID: \"dcf21347-4c2f-4d66-9137-c24ee63a31de\") " pod="hostpath-provisioner/csi-hostpathplugin-7jgjj" Feb 18 14:57:24 crc kubenswrapper[4896]: E0218 14:57:24.432104 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:24.932092756 +0000 UTC m=+143.861427039 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.432317 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2573759f-2db1-452a-b310-f2a170099ec6-config\") pod \"service-ca-operator-777779d784-rx8f2\" (UID: \"2573759f-2db1-452a-b310-f2a170099ec6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-rx8f2" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.432997 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4122c286-47e3-460d-bd25-323012847f12-config\") pod \"kube-controller-manager-operator-78b949d7b-2bncs\" (UID: \"4122c286-47e3-460d-bd25-323012847f12\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bncs" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.433925 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/64e32b45-d067-4151-980d-36ea8f3116e9-srv-cert\") pod \"catalog-operator-68c6474976-66zwg\" (UID: \"64e32b45-d067-4151-980d-36ea8f3116e9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-66zwg" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.435668 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6225d3ae-2a4e-479c-aa70-6fdbd2ee488b-etcd-client\") pod \"etcd-operator-b45778765-fhpqt\" (UID: \"6225d3ae-2a4e-479c-aa70-6fdbd2ee488b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fhpqt" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.436740 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b60354fe-e785-4e88-af1c-0d846d9ff59a-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bjbsw\" (UID: \"b60354fe-e785-4e88-af1c-0d846d9ff59a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bjbsw" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.436719 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/dcf21347-4c2f-4d66-9137-c24ee63a31de-csi-data-dir\") pod \"csi-hostpathplugin-7jgjj\" (UID: \"dcf21347-4c2f-4d66-9137-c24ee63a31de\") " pod="hostpath-provisioner/csi-hostpathplugin-7jgjj" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.436980 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/dcf21347-4c2f-4d66-9137-c24ee63a31de-plugins-dir\") pod \"csi-hostpathplugin-7jgjj\" (UID: \"dcf21347-4c2f-4d66-9137-c24ee63a31de\") " pod="hostpath-provisioner/csi-hostpathplugin-7jgjj" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.438024 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b0388f1f-8b9c-46e4-a7f0-1fad187d6d65-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-v59s8\" (UID: \"b0388f1f-8b9c-46e4-a7f0-1fad187d6d65\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v59s8" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.438610 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/7b9d6a43-e957-4d02-9928-020f9f17f7a9-tmpfs\") pod \"packageserver-d55dfcdfc-l8fcf\" (UID: \"7b9d6a43-e957-4d02-9928-020f9f17f7a9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l8fcf" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.439811 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/013d2ce9-60ac-4abf-945d-3340ffa55b72-node-bootstrap-token\") pod \"machine-config-server-s95mb\" (UID: \"013d2ce9-60ac-4abf-945d-3340ffa55b72\") " pod="openshift-machine-config-operator/machine-config-server-s95mb" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.440007 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0f95dce4-6d68-44ac-9dd3-215664bb1d75-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xrhcj\" (UID: \"0f95dce4-6d68-44ac-9dd3-215664bb1d75\") " pod="openshift-marketplace/marketplace-operator-79b997595-xrhcj" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.440191 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4122c286-47e3-460d-bd25-323012847f12-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-2bncs\" (UID: \"4122c286-47e3-460d-bd25-323012847f12\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bncs" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.441400 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/8db0982f-0e1d-493c-8b32-64190372fcb9-signing-key\") pod \"service-ca-9c57cc56f-ccl77\" (UID: \"8db0982f-0e1d-493c-8b32-64190372fcb9\") " pod="openshift-service-ca/service-ca-9c57cc56f-ccl77" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.441422 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/013d2ce9-60ac-4abf-945d-3340ffa55b72-certs\") pod \"machine-config-server-s95mb\" (UID: \"013d2ce9-60ac-4abf-945d-3340ffa55b72\") " pod="openshift-machine-config-operator/machine-config-server-s95mb" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.441487 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/dcf21347-4c2f-4d66-9137-c24ee63a31de-mountpoint-dir\") pod \"csi-hostpathplugin-7jgjj\" (UID: \"dcf21347-4c2f-4d66-9137-c24ee63a31de\") " pod="hostpath-provisioner/csi-hostpathplugin-7jgjj" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.441582 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxhp2\" (UniqueName: \"kubernetes.io/projected/f8abcf75-9577-4dfe-a252-40aff339ffce-kube-api-access-gxhp2\") pod \"openshift-controller-manager-operator-756b6f6bc6-gqb6w\" (UID: \"f8abcf75-9577-4dfe-a252-40aff339ffce\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gqb6w" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.442921 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/8db0982f-0e1d-493c-8b32-64190372fcb9-signing-cabundle\") pod \"service-ca-9c57cc56f-ccl77\" (UID: \"8db0982f-0e1d-493c-8b32-64190372fcb9\") " pod="openshift-service-ca/service-ca-9c57cc56f-ccl77" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.443908 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dd10c491-df5d-43a1-961d-613e2a90be23-config-volume\") pod \"collect-profiles-29523765-frmvz\" (UID: \"dd10c491-df5d-43a1-961d-613e2a90be23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523765-frmvz" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.445073 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8fa6844e-65a5-401e-b329-2a49dc5debe4-cert\") pod \"ingress-canary-znxvs\" (UID: \"8fa6844e-65a5-401e-b329-2a49dc5debe4\") " pod="openshift-ingress-canary/ingress-canary-znxvs" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.445739 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0f95dce4-6d68-44ac-9dd3-215664bb1d75-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xrhcj\" (UID: \"0f95dce4-6d68-44ac-9dd3-215664bb1d75\") " pod="openshift-marketplace/marketplace-operator-79b997595-xrhcj" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.445873 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/8c1a0565-e55b-410e-9b63-12ababdb69eb-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-jpg48\" (UID: \"8c1a0565-e55b-410e-9b63-12ababdb69eb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jpg48" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.446000 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/86b5fc8f-1b76-43ad-b283-0d46697be7aa-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hv6vt\" (UID: \"86b5fc8f-1b76-43ad-b283-0d46697be7aa\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hv6vt" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.446344 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/64e32b45-d067-4151-980d-36ea8f3116e9-profile-collector-cert\") pod \"catalog-operator-68c6474976-66zwg\" (UID: \"64e32b45-d067-4151-980d-36ea8f3116e9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-66zwg" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.450440 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7b9d6a43-e957-4d02-9928-020f9f17f7a9-apiservice-cert\") pod \"packageserver-d55dfcdfc-l8fcf\" (UID: \"7b9d6a43-e957-4d02-9928-020f9f17f7a9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l8fcf" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.450558 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9feb8b3f-6359-41f1-8648-c827ce9c5ec4-metrics-tls\") pod \"dns-default-5795c\" (UID: \"9feb8b3f-6359-41f1-8648-c827ce9c5ec4\") " pod="openshift-dns/dns-default-5795c" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.451200 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6225d3ae-2a4e-479c-aa70-6fdbd2ee488b-serving-cert\") pod \"etcd-operator-b45778765-fhpqt\" (UID: \"6225d3ae-2a4e-479c-aa70-6fdbd2ee488b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fhpqt" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.451437 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dd10c491-df5d-43a1-961d-613e2a90be23-secret-volume\") pod \"collect-profiles-29523765-frmvz\" (UID: \"dd10c491-df5d-43a1-961d-613e2a90be23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523765-frmvz" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.451595 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7b9d6a43-e957-4d02-9928-020f9f17f7a9-webhook-cert\") pod \"packageserver-d55dfcdfc-l8fcf\" (UID: \"7b9d6a43-e957-4d02-9928-020f9f17f7a9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l8fcf" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.451677 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2573759f-2db1-452a-b310-f2a170099ec6-serving-cert\") pod \"service-ca-operator-777779d784-rx8f2\" (UID: \"2573759f-2db1-452a-b310-f2a170099ec6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-rx8f2" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.452159 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b60354fe-e785-4e88-af1c-0d846d9ff59a-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bjbsw\" (UID: \"b60354fe-e785-4e88-af1c-0d846d9ff59a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bjbsw" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.452738 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/5c0e215f-d940-415a-90d3-f5faf3e537de-profile-collector-cert\") pod \"olm-operator-6b444d44fb-j99bk\" (UID: \"5c0e215f-d940-415a-90d3-f5faf3e537de\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j99bk" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.453317 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b0388f1f-8b9c-46e4-a7f0-1fad187d6d65-proxy-tls\") pod \"machine-config-controller-84d6567774-v59s8\" (UID: \"b0388f1f-8b9c-46e4-a7f0-1fad187d6d65\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v59s8" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.457664 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/5c0e215f-d940-415a-90d3-f5faf3e537de-srv-cert\") pod \"olm-operator-6b444d44fb-j99bk\" (UID: \"5c0e215f-d940-415a-90d3-f5faf3e537de\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j99bk" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.469830 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdlmm\" (UniqueName: \"kubernetes.io/projected/710fc779-629b-4d37-a820-e08f5289677f-kube-api-access-mdlmm\") pod \"apiserver-76f77b778f-tgwt9\" (UID: \"710fc779-629b-4d37-a820-e08f5289677f\") " pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.484067 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-n25dz"] Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.491193 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6pk7\" (UniqueName: \"kubernetes.io/projected/76cb1e0e-4321-4450-9fcb-b505a7a1d969-kube-api-access-q6pk7\") pod \"oauth-openshift-558db77b4-shvc6\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.516790 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r28rq\" (UniqueName: \"kubernetes.io/projected/3cd5b209-b64c-4ce1-85e2-e00e568b4bbe-kube-api-access-r28rq\") pod \"cluster-samples-operator-665b6dd947-bw75d\" (UID: \"3cd5b209-b64c-4ce1-85e2-e00e568b4bbe\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bw75d" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.523041 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bw75d" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.532442 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:24 crc kubenswrapper[4896]: E0218 14:57:24.532905 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:25.032889473 +0000 UTC m=+143.962223686 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.534415 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6pnr\" (UniqueName: \"kubernetes.io/projected/21fb3328-3882-4644-9a94-b7949f48db54-kube-api-access-j6pnr\") pod \"machine-config-operator-74547568cd-6frkl\" (UID: \"21fb3328-3882-4644-9a94-b7949f48db54\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6frkl" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.554113 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.554752 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhgdz"] Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.557987 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tb9ks\" (UniqueName: \"kubernetes.io/projected/918ef2a1-8828-4b17-b30a-de957f009104-kube-api-access-tb9ks\") pod \"multus-admission-controller-857f4d67dd-z7xdz\" (UID: \"918ef2a1-8828-4b17-b30a-de957f009104\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-z7xdz" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.570823 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gqb6w" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.572649 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75xt2\" (UniqueName: \"kubernetes.io/projected/71145266-38d8-495b-83c2-cff3c68e705d-kube-api-access-75xt2\") pod \"openshift-apiserver-operator-796bbdcf4f-2ftds\" (UID: \"71145266-38d8-495b-83c2-cff3c68e705d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ftds" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.578950 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mcjhx" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.593086 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-272j4\" (UniqueName: \"kubernetes.io/projected/31b0687b-5c42-4259-b96d-284c1aa487b4-kube-api-access-272j4\") pod \"controller-manager-879f6c89f-27bbf\" (UID: \"31b0687b-5c42-4259-b96d-284c1aa487b4\") " pod="openshift-controller-manager/controller-manager-879f6c89f-27bbf" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.645923 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-z7xdz" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.645988 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6frkl" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.646474 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.647095 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.648016 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fjmq5"] Feb 18 14:57:24 crc kubenswrapper[4896]: E0218 14:57:24.649170 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:25.149149778 +0000 UTC m=+144.078483991 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.651270 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t2wdv" event={"ID":"f2da710c-c8cd-4e5f-8ba3-045858050ade","Type":"ContainerStarted","Data":"10ed7fcbfeae461d0265c35743df07998d657285bd7ad2065b156d6f140f0492"} Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.651304 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t2wdv" event={"ID":"f2da710c-c8cd-4e5f-8ba3-045858050ade","Type":"ContainerStarted","Data":"465f02d9f2f1dfd10c9873fd77bccd59729325064cea98da9578fa7128dabea7"} Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.660045 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bjgv\" (UniqueName: \"kubernetes.io/projected/036e7602-7608-4033-b18d-d859a7eddebd-kube-api-access-4bjgv\") pod \"downloads-7954f5f757-6v4cn\" (UID: \"036e7602-7608-4033-b18d-d859a7eddebd\") " pod="openshift-console/downloads-7954f5f757-6v4cn" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.661062 4896 generic.go:334] "Generic (PLEG): container finished" podID="81190a62-b874-433f-8e50-f6dfe001f6d2" containerID="56aaf93401579a201e58635284508c6a25e5ed0d97a7fdd61a79875c979ed687" exitCode=0 Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.661127 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" event={"ID":"81190a62-b874-433f-8e50-f6dfe001f6d2","Type":"ContainerDied","Data":"56aaf93401579a201e58635284508c6a25e5ed0d97a7fdd61a79875c979ed687"} Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.661156 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" event={"ID":"81190a62-b874-433f-8e50-f6dfe001f6d2","Type":"ContainerStarted","Data":"816b74a23e678c566392d85d311022ebece46122e27cefb801bba6dc2b0231dd"} Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.663458 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dlghr" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.664176 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhgdz" event={"ID":"30e5171d-473c-4388-977d-95ce80ae1b79","Type":"ContainerStarted","Data":"4d5e031f01ab30e991f1d2cd71fc745be0380a20ba331ef29d18f4e5f09e75e4"} Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.665533 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v6thv" event={"ID":"0c6cb779-d660-45f8-8fb2-5829d42ee030","Type":"ContainerStarted","Data":"48efb299ab8ed41e49575c9749f3da5efa5277019eabe5db6fe02dcf8d920bd0"} Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.665587 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v6thv" event={"ID":"0c6cb779-d660-45f8-8fb2-5829d42ee030","Type":"ContainerStarted","Data":"e1bcb0789d6c13f653780242f8448b2b193c29cf60ec5b474aead9d95a211a14"} Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.666527 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v6thv" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.670173 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-n25dz" event={"ID":"fd2f0b92-9f4d-4c2c-b83f-7530706cfa09","Type":"ContainerStarted","Data":"97c019f38d3affa35fe62f0587c4986a8e2f5ddfef050da4b9cb8223c35c2eb0"} Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.672757 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-8cq46" event={"ID":"7a51266e-326e-4b5c-a25a-fb61bf1f3757","Type":"ContainerStarted","Data":"d1373ec24b187c32a516b7ab25b4c27a68be6c32378507dd3a27d9e736b67632"} Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.672795 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-8cq46" event={"ID":"7a51266e-326e-4b5c-a25a-fb61bf1f3757","Type":"ContainerStarted","Data":"9071936d0faf6e8ddcc4832ae2d941a12b602f63bb32ba26985268b3ae6ceae8"} Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.673833 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f88rx\" (UniqueName: \"kubernetes.io/projected/351aa26e-939a-4b69-a7ee-55be7111ee6e-kube-api-access-f88rx\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.675943 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-22f85" event={"ID":"42ba2ce9-d014-4090-b4b2-21a2ae8cea4c","Type":"ContainerStarted","Data":"4aa9fbd4a2c3bf0cf99d596bcfb1c2fa0b92d047fae572a4aa26e72ef653c33e"} Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.676105 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-22f85" event={"ID":"42ba2ce9-d014-4090-b4b2-21a2ae8cea4c","Type":"ContainerStarted","Data":"c82303d4612d31e1e1ed5fb14cb76b51e5e8c2f493f164198c82f006a05a85ca"} Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.676157 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-22f85" event={"ID":"42ba2ce9-d014-4090-b4b2-21a2ae8cea4c","Type":"ContainerStarted","Data":"ccc27c196c5c79e73d0a224e38528fcd562c015e3a8169e64a01038ff9c377b1"} Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.683704 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlg8k\" (UniqueName: \"kubernetes.io/projected/fa6c59b4-6823-47ef-a8a6-52c94e53ef4c-kube-api-access-wlg8k\") pod \"console-f9d7485db-4bxkk\" (UID: \"fa6c59b4-6823-47ef-a8a6-52c94e53ef4c\") " pod="openshift-console/console-f9d7485db-4bxkk" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.689081 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqzfn\" (UniqueName: \"kubernetes.io/projected/17344065-a988-48bb-aff9-d51ffd76f54b-kube-api-access-cqzfn\") pod \"dns-operator-744455d44c-rmc7c\" (UID: \"17344065-a988-48bb-aff9-d51ffd76f54b\") " pod="openshift-dns-operator/dns-operator-744455d44c-rmc7c" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.708111 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xf69f\" (UniqueName: \"kubernetes.io/projected/8fa6844e-65a5-401e-b329-2a49dc5debe4-kube-api-access-xf69f\") pod \"ingress-canary-znxvs\" (UID: \"8fa6844e-65a5-401e-b329-2a49dc5debe4\") " pod="openshift-ingress-canary/ingress-canary-znxvs" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.723716 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v6thv" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.736571 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txk8c\" (UniqueName: \"kubernetes.io/projected/6225d3ae-2a4e-479c-aa70-6fdbd2ee488b-kube-api-access-txk8c\") pod \"etcd-operator-b45778765-fhpqt\" (UID: \"6225d3ae-2a4e-479c-aa70-6fdbd2ee488b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fhpqt" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.739748 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7l5m5\" (UniqueName: \"kubernetes.io/projected/64e32b45-d067-4151-980d-36ea8f3116e9-kube-api-access-7l5m5\") pod \"catalog-operator-68c6474976-66zwg\" (UID: \"64e32b45-d067-4151-980d-36ea8f3116e9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-66zwg" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.748203 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:24 crc kubenswrapper[4896]: E0218 14:57:24.750017 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:25.249998826 +0000 UTC m=+144.179333039 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.754761 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-6v4cn" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.757192 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r9vl\" (UniqueName: \"kubernetes.io/projected/9feb8b3f-6359-41f1-8648-c827ce9c5ec4-kube-api-access-4r9vl\") pod \"dns-default-5795c\" (UID: \"9feb8b3f-6359-41f1-8648-c827ce9c5ec4\") " pod="openshift-dns/dns-default-5795c" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.758913 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-966wr"] Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.760642 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-p7zgl"] Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.771607 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-5795c" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.777331 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b82p9\" (UniqueName: \"kubernetes.io/projected/8db0982f-0e1d-493c-8b32-64190372fcb9-kube-api-access-b82p9\") pod \"service-ca-9c57cc56f-ccl77\" (UID: \"8db0982f-0e1d-493c-8b32-64190372fcb9\") " pod="openshift-service-ca/service-ca-9c57cc56f-ccl77" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.792640 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99b5p\" (UniqueName: \"kubernetes.io/projected/013d2ce9-60ac-4abf-945d-3340ffa55b72-kube-api-access-99b5p\") pod \"machine-config-server-s95mb\" (UID: \"013d2ce9-60ac-4abf-945d-3340ffa55b72\") " pod="openshift-machine-config-operator/machine-config-server-s95mb" Feb 18 14:57:24 crc kubenswrapper[4896]: W0218 14:57:24.795862 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7f8cbed1_6086_4fb6_bab5_0563bbcb751c.slice/crio-05b59a99f5c9c8beeb3a17361546f7bc2cdb29a65233866f36eacf87a9b71c72 WatchSource:0}: Error finding container 05b59a99f5c9c8beeb3a17361546f7bc2cdb29a65233866f36eacf87a9b71c72: Status 404 returned error can't find the container with id 05b59a99f5c9c8beeb3a17361546f7bc2cdb29a65233866f36eacf87a9b71c72 Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.805949 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-znxvs" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.806848 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ftds" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.811709 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tq74b\" (UniqueName: \"kubernetes.io/projected/dd10c491-df5d-43a1-961d-613e2a90be23-kube-api-access-tq74b\") pod \"collect-profiles-29523765-frmvz\" (UID: \"dd10c491-df5d-43a1-961d-613e2a90be23\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523765-frmvz" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.831408 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2jpm\" (UniqueName: \"kubernetes.io/projected/5c0e215f-d940-415a-90d3-f5faf3e537de-kube-api-access-j2jpm\") pod \"olm-operator-6b444d44fb-j99bk\" (UID: \"5c0e215f-d940-415a-90d3-f5faf3e537de\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j99bk" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.835746 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bw75d"] Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.844217 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-27bbf" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.850441 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:24 crc kubenswrapper[4896]: E0218 14:57:24.850908 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:25.350893815 +0000 UTC m=+144.280228028 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.859308 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwt8v\" (UniqueName: \"kubernetes.io/projected/0f95dce4-6d68-44ac-9dd3-215664bb1d75-kube-api-access-lwt8v\") pod \"marketplace-operator-79b997595-xrhcj\" (UID: \"0f95dce4-6d68-44ac-9dd3-215664bb1d75\") " pod="openshift-marketplace/marketplace-operator-79b997595-xrhcj" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.871712 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rbcf\" (UniqueName: \"kubernetes.io/projected/86b5fc8f-1b76-43ad-b283-0d46697be7aa-kube-api-access-5rbcf\") pod \"package-server-manager-789f6589d5-hv6vt\" (UID: \"86b5fc8f-1b76-43ad-b283-0d46697be7aa\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hv6vt" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.902983 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gqb6w"] Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.909374 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-4bxkk" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.927363 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b60354fe-e785-4e88-af1c-0d846d9ff59a-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bjbsw\" (UID: \"b60354fe-e785-4e88-af1c-0d846d9ff59a\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bjbsw" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.930294 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-rmc7c" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.935783 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-mcjhx"] Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.939362 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xczsj\" (UniqueName: \"kubernetes.io/projected/dcf21347-4c2f-4d66-9137-c24ee63a31de-kube-api-access-xczsj\") pod \"csi-hostpathplugin-7jgjj\" (UID: \"dcf21347-4c2f-4d66-9137-c24ee63a31de\") " pod="hostpath-provisioner/csi-hostpathplugin-7jgjj" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.948165 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4122c286-47e3-460d-bd25-323012847f12-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-2bncs\" (UID: \"4122c286-47e3-460d-bd25-323012847f12\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bncs" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.951693 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:24 crc kubenswrapper[4896]: E0218 14:57:24.951855 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:25.451827336 +0000 UTC m=+144.381161549 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.951903 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:24 crc kubenswrapper[4896]: E0218 14:57:24.953131 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:25.45265911 +0000 UTC m=+144.381993323 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.970671 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xrhcj" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.974927 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q5kn\" (UniqueName: \"kubernetes.io/projected/2573759f-2db1-452a-b310-f2a170099ec6-kube-api-access-8q5kn\") pod \"service-ca-operator-777779d784-rx8f2\" (UID: \"2573759f-2db1-452a-b310-f2a170099ec6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-rx8f2" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.976359 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqvw5\" (UniqueName: \"kubernetes.io/projected/7b9d6a43-e957-4d02-9928-020f9f17f7a9-kube-api-access-lqvw5\") pod \"packageserver-d55dfcdfc-l8fcf\" (UID: \"7b9d6a43-e957-4d02-9928-020f9f17f7a9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l8fcf" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.987688 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-fhpqt" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.989587 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bncs" Feb 18 14:57:24 crc kubenswrapper[4896]: I0218 14:57:24.996395 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bjbsw" Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.002714 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hv6vt" Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.016464 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-ccl77" Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.017991 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jnkn\" (UniqueName: \"kubernetes.io/projected/b0388f1f-8b9c-46e4-a7f0-1fad187d6d65-kube-api-access-5jnkn\") pod \"machine-config-controller-84d6567774-v59s8\" (UID: \"b0388f1f-8b9c-46e4-a7f0-1fad187d6d65\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v59s8" Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.029384 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523765-frmvz" Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.029667 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-rx8f2" Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.031904 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ck9fl\" (UniqueName: \"kubernetes.io/projected/8c1a0565-e55b-410e-9b63-12ababdb69eb-kube-api-access-ck9fl\") pod \"control-plane-machine-set-operator-78cbb6b69f-jpg48\" (UID: \"8c1a0565-e55b-410e-9b63-12ababdb69eb\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jpg48" Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.033072 4896 csr.go:261] certificate signing request csr-wl75s is approved, waiting to be issued Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.037549 4896 csr.go:257] certificate signing request csr-wl75s is issued Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.039557 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgmr5\" (UniqueName: \"kubernetes.io/projected/0e9fc2ab-6d3d-441d-8591-3d64ce77c62d-kube-api-access-pgmr5\") pod \"migrator-59844c95c7-pclrg\" (UID: \"0e9fc2ab-6d3d-441d-8591-3d64ce77c62d\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pclrg" Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.039850 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-66zwg" Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.043346 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j99bk" Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.050565 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l8fcf" Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.052877 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:25 crc kubenswrapper[4896]: E0218 14:57:25.053352 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:25.553333563 +0000 UTC m=+144.482667776 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.062061 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-s95mb" Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.063739 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v59s8" Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.099791 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-7jgjj" Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.110192 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-tgwt9"] Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.154726 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:25 crc kubenswrapper[4896]: E0218 14:57:25.155039 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:25.655028335 +0000 UTC m=+144.584362548 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.163879 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-6frkl"] Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.255509 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:25 crc kubenswrapper[4896]: E0218 14:57:25.255669 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:25.755643206 +0000 UTC m=+144.684977419 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.255802 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:25 crc kubenswrapper[4896]: E0218 14:57:25.256147 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:25.75613083 +0000 UTC m=+144.685465043 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:25 crc kubenswrapper[4896]: W0218 14:57:25.266820 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod710fc779_629b_4d37_a820_e08f5289677f.slice/crio-2304c02517601fd29e60b56669b8cc736b7e510a947adea1645bb585d729d779 WatchSource:0}: Error finding container 2304c02517601fd29e60b56669b8cc736b7e510a947adea1645bb585d729d779: Status 404 returned error can't find the container with id 2304c02517601fd29e60b56669b8cc736b7e510a947adea1645bb585d729d779 Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.272098 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-dlghr"] Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.275600 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pclrg" Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.289739 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-shvc6"] Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.309317 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jpg48" Feb 18 14:57:25 crc kubenswrapper[4896]: W0218 14:57:25.332174 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21fb3328_3882_4644_9a94_b7949f48db54.slice/crio-4ad29181627b3c9e40c57d5feccb27e87a3285fae0ad0516e03bc94fa8cfa24e WatchSource:0}: Error finding container 4ad29181627b3c9e40c57d5feccb27e87a3285fae0ad0516e03bc94fa8cfa24e: Status 404 returned error can't find the container with id 4ad29181627b3c9e40c57d5feccb27e87a3285fae0ad0516e03bc94fa8cfa24e Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.338230 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-8cq46" Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.351000 4896 patch_prober.go:28] interesting pod/router-default-5444994796-8cq46 container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.351054 4896 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8cq46" podUID="7a51266e-326e-4b5c-a25a-fb61bf1f3757" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.359041 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:25 crc kubenswrapper[4896]: E0218 14:57:25.359314 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:25.859299326 +0000 UTC m=+144.788633529 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.374778 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-z7xdz"] Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.423430 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-znxvs"] Feb 18 14:57:25 crc kubenswrapper[4896]: W0218 14:57:25.445410 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod918ef2a1_8828_4b17_b30a_de957f009104.slice/crio-9607de89b9015b097141d5fddbbd441d574177cdcacde75c6f5a4ad457f81e7f WatchSource:0}: Error finding container 9607de89b9015b097141d5fddbbd441d574177cdcacde75c6f5a4ad457f81e7f: Status 404 returned error can't find the container with id 9607de89b9015b097141d5fddbbd441d574177cdcacde75c6f5a4ad457f81e7f Feb 18 14:57:25 crc kubenswrapper[4896]: W0218 14:57:25.450713 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod013d2ce9_60ac_4abf_945d_3340ffa55b72.slice/crio-06748f04aa313b27d961aa470c883466a4826c9fcec6d1a9bfe1c08f2b90ffa5 WatchSource:0}: Error finding container 06748f04aa313b27d961aa470c883466a4826c9fcec6d1a9bfe1c08f2b90ffa5: Status 404 returned error can't find the container with id 06748f04aa313b27d961aa470c883466a4826c9fcec6d1a9bfe1c08f2b90ffa5 Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.460147 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:25 crc kubenswrapper[4896]: E0218 14:57:25.460501 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:25.960485633 +0000 UTC m=+144.889819846 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.522777 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-6v4cn"] Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.525678 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-5795c"] Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.561501 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:25 crc kubenswrapper[4896]: E0218 14:57:25.561904 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:26.061888307 +0000 UTC m=+144.991222520 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.663261 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:25 crc kubenswrapper[4896]: E0218 14:57:25.663596 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:26.163573169 +0000 UTC m=+145.092907382 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.690742 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-27bbf"] Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.697363 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ftds"] Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.710621 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-6v4cn" event={"ID":"036e7602-7608-4033-b18d-d859a7eddebd","Type":"ContainerStarted","Data":"736c5b1ad2262285a3ff9aca662c2e7b78ce0c7eb266d4b171b6ee65f5192ed0"} Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.712637 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-s95mb" event={"ID":"013d2ce9-60ac-4abf-945d-3340ffa55b72","Type":"ContainerStarted","Data":"06748f04aa313b27d961aa470c883466a4826c9fcec6d1a9bfe1c08f2b90ffa5"} Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.719471 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhgdz" event={"ID":"30e5171d-473c-4388-977d-95ce80ae1b79","Type":"ContainerStarted","Data":"8e6c76ec1056f5390d8c1875ad27ed4c9b37286754de952c3af445b297c66142"} Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.721056 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-4bxkk"] Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.721627 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bncs"] Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.723938 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" event={"ID":"81190a62-b874-433f-8e50-f6dfe001f6d2","Type":"ContainerStarted","Data":"44ddc76905ae37e5b973be74a11e8acfef1001b98e055f61be0542b3328aec50"} Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.726900 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-p7zgl" event={"ID":"a1659f4f-bd6e-4d58-a56b-69b14ab57ed0","Type":"ContainerStarted","Data":"c34e487432431d80e1a48af291bcf98c4f60d1a0d1b2558b14f96ba208f3ea95"} Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.726929 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-p7zgl" event={"ID":"a1659f4f-bd6e-4d58-a56b-69b14ab57ed0","Type":"ContainerStarted","Data":"c8833e1dd2573bac9504c1c9038c32a62f85b46b31c1db61d8fdf13105678afb"} Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.727182 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-p7zgl" Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.728209 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-5795c" event={"ID":"9feb8b3f-6359-41f1-8648-c827ce9c5ec4","Type":"ContainerStarted","Data":"a387cf439167fcb838bc2ef4788d074a03ae4b40bae73462c06a832912916065"} Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.732455 4896 patch_prober.go:28] interesting pod/console-operator-58897d9998-p7zgl container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.732497 4896 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-p7zgl" podUID="a1659f4f-bd6e-4d58-a56b-69b14ab57ed0" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.732977 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dlghr" event={"ID":"37bf7cb8-e127-458a-ab07-efb2c111ce02","Type":"ContainerStarted","Data":"92e0fb4dc0f33b7195fcbbb05724f3b636a183c774429618e76e46c1c1754f84"} Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.734807 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-966wr" event={"ID":"7f8cbed1-6086-4fb6-bab5-0563bbcb751c","Type":"ContainerStarted","Data":"e0de7e79cd7260b81cd1d7aee6dce950cc13cddc1b6991309e8f904082f5cfce"} Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.734834 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-966wr" event={"ID":"7f8cbed1-6086-4fb6-bab5-0563bbcb751c","Type":"ContainerStarted","Data":"05b59a99f5c9c8beeb3a17361546f7bc2cdb29a65233866f36eacf87a9b71c72"} Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.738583 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-z7xdz" event={"ID":"918ef2a1-8828-4b17-b30a-de957f009104","Type":"ContainerStarted","Data":"9607de89b9015b097141d5fddbbd441d574177cdcacde75c6f5a4ad457f81e7f"} Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.742728 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" event={"ID":"710fc779-629b-4d37-a820-e08f5289677f","Type":"ContainerStarted","Data":"2304c02517601fd29e60b56669b8cc736b7e510a947adea1645bb585d729d779"} Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.745109 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-n25dz" event={"ID":"fd2f0b92-9f4d-4c2c-b83f-7530706cfa09","Type":"ContainerStarted","Data":"50541756d47dab3623a35ee556814803c7ab99e606bc55e1ae678202c873b0d3"} Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.747403 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mcjhx" event={"ID":"23b10824-dc70-459f-b037-6f8f00ae6a04","Type":"ContainerStarted","Data":"c3e10f8540686c5f670b5856ce6e9def5128dcd882c3fcf8d4e73f5c8ca74a09"} Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.755666 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fjmq5" event={"ID":"abcc561e-04a7-4cf3-a5d3-f2937d0f18c0","Type":"ContainerStarted","Data":"f31a98133a5ccffc03ee4233c9c7024812946d2c2faaa4609e49d68ad4ea05f0"} Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.755725 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fjmq5" event={"ID":"abcc561e-04a7-4cf3-a5d3-f2937d0f18c0","Type":"ContainerStarted","Data":"9623342142a01b70dc9b497c62f789c96ae6b94a76b625640df3015c72a3eb5c"} Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.757530 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gqb6w" event={"ID":"f8abcf75-9577-4dfe-a252-40aff339ffce","Type":"ContainerStarted","Data":"b50cc0757d8ce39b1e92e9186fc05789b9a9a25e262377d71e98ccd1347c9048"} Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.758419 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bw75d" event={"ID":"3cd5b209-b64c-4ce1-85e2-e00e568b4bbe","Type":"ContainerStarted","Data":"d449e9150459430f2de767d686e05d14bdafb8a693a6aba36faa7c70b0d4c1f2"} Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.758444 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bw75d" event={"ID":"3cd5b209-b64c-4ce1-85e2-e00e568b4bbe","Type":"ContainerStarted","Data":"587667d321e93e5106f05b9f3038c6ce6340d82daeca0b47bb60a9f089fd4e60"} Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.759104 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-znxvs" event={"ID":"8fa6844e-65a5-401e-b329-2a49dc5debe4","Type":"ContainerStarted","Data":"eaeaf29d366613158e6b2f82960250368bb7224b91bd77b23ba144734f7c84ac"} Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.761304 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" event={"ID":"76cb1e0e-4321-4450-9fcb-b505a7a1d969","Type":"ContainerStarted","Data":"a46c46add4951b8d35cc1f7cd2f6966369f1821f641e11c09321c0d99de9be14"} Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.764400 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:25 crc kubenswrapper[4896]: E0218 14:57:25.764868 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:26.264848289 +0000 UTC m=+145.194182502 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.765282 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6frkl" event={"ID":"21fb3328-3882-4644-9a94-b7949f48db54","Type":"ContainerStarted","Data":"4ad29181627b3c9e40c57d5feccb27e87a3285fae0ad0516e03bc94fa8cfa24e"} Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.777942 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-fhpqt"] Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.791661 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-rmc7c"] Feb 18 14:57:25 crc kubenswrapper[4896]: E0218 14:57:25.867175 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:26.367159339 +0000 UTC m=+145.296493552 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.866870 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.914289 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-t2wdv" podStartSLOduration=121.914266937 podStartE2EDuration="2m1.914266937s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:25.902932248 +0000 UTC m=+144.832266461" watchObservedRunningTime="2026-02-18 14:57:25.914266937 +0000 UTC m=+144.843601150" Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.970942 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:25 crc kubenswrapper[4896]: E0218 14:57:25.971167 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:26.471146818 +0000 UTC m=+145.400481031 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:25 crc kubenswrapper[4896]: I0218 14:57:25.972625 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:25 crc kubenswrapper[4896]: E0218 14:57:25.973582 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:26.473564458 +0000 UTC m=+145.402898671 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.029597 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l8fcf"] Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.040438 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-18 14:52:25 +0000 UTC, rotation deadline is 2026-11-09 14:59:42.822156114 +0000 UTC Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.040490 4896 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6336h2m16.781669033s for next certificate rotation Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.077264 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:26 crc kubenswrapper[4896]: E0218 14:57:26.077565 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:26.577547547 +0000 UTC m=+145.506881760 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.161531 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-8cq46" podStartSLOduration=122.161509015 podStartE2EDuration="2m2.161509015s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:26.159148767 +0000 UTC m=+145.088482980" watchObservedRunningTime="2026-02-18 14:57:26.161509015 +0000 UTC m=+145.090843238" Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.178292 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:26 crc kubenswrapper[4896]: E0218 14:57:26.178605 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:26.678593361 +0000 UTC m=+145.607927574 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.265633 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-22f85" podStartSLOduration=122.265614668 podStartE2EDuration="2m2.265614668s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:26.264630099 +0000 UTC m=+145.193964322" watchObservedRunningTime="2026-02-18 14:57:26.265614668 +0000 UTC m=+145.194948881" Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.280385 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:26 crc kubenswrapper[4896]: E0218 14:57:26.280760 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:26.780739717 +0000 UTC m=+145.710073940 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.345282 4896 patch_prober.go:28] interesting pod/router-default-5444994796-8cq46 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 14:57:26 crc kubenswrapper[4896]: [-]has-synced failed: reason withheld Feb 18 14:57:26 crc kubenswrapper[4896]: [+]process-running ok Feb 18 14:57:26 crc kubenswrapper[4896]: healthz check failed Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.345338 4896 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8cq46" podUID="7a51266e-326e-4b5c-a25a-fb61bf1f3757" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.381925 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.382219 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xrhcj"] Feb 18 14:57:26 crc kubenswrapper[4896]: E0218 14:57:26.382406 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:26.882381008 +0000 UTC m=+145.811715211 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.410340 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-ccl77"] Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.415053 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-v59s8"] Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.425703 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bjbsw"] Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.425846 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j99bk"] Feb 18 14:57:26 crc kubenswrapper[4896]: W0218 14:57:26.478893 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8db0982f_0e1d_493c_8b32_64190372fcb9.slice/crio-e3c4a9d03bc8b8969a13d5556e3c510019742807f001e8ff78d08aa1da6250c3 WatchSource:0}: Error finding container e3c4a9d03bc8b8969a13d5556e3c510019742807f001e8ff78d08aa1da6250c3: Status 404 returned error can't find the container with id e3c4a9d03bc8b8969a13d5556e3c510019742807f001e8ff78d08aa1da6250c3 Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.480576 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-rx8f2"] Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.482752 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:26 crc kubenswrapper[4896]: E0218 14:57:26.484769 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:26.98474454 +0000 UTC m=+145.914078763 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:26 crc kubenswrapper[4896]: W0218 14:57:26.493497 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb60354fe_e785_4e88_af1c_0d846d9ff59a.slice/crio-59a1c0311f0c9a8fd3b60107c03b52d4f44ac7d46dbd8f3455070284cf639101 WatchSource:0}: Error finding container 59a1c0311f0c9a8fd3b60107c03b52d4f44ac7d46dbd8f3455070284cf639101: Status 404 returned error can't find the container with id 59a1c0311f0c9a8fd3b60107c03b52d4f44ac7d46dbd8f3455070284cf639101 Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.514473 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hv6vt"] Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.547186 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v6thv" podStartSLOduration=122.547172702 podStartE2EDuration="2m2.547172702s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:26.546827212 +0000 UTC m=+145.476161425" watchObservedRunningTime="2026-02-18 14:57:26.547172702 +0000 UTC m=+145.476506915" Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.567698 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-66zwg"] Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.584344 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-pclrg"] Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.585525 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:26 crc kubenswrapper[4896]: E0218 14:57:26.585841 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:27.085826754 +0000 UTC m=+146.015160967 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.587105 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jpg48"] Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.594875 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523765-frmvz"] Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.597509 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-7jgjj"] Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.687190 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:26 crc kubenswrapper[4896]: E0218 14:57:26.688225 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:27.188185426 +0000 UTC m=+146.117519689 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.688523 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:26 crc kubenswrapper[4896]: E0218 14:57:26.690023 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:27.190006639 +0000 UTC m=+146.119340852 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.723395 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-p7zgl" podStartSLOduration=122.723378148 podStartE2EDuration="2m2.723378148s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:26.686027883 +0000 UTC m=+145.615362096" watchObservedRunningTime="2026-02-18 14:57:26.723378148 +0000 UTC m=+145.652712361" Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.769301 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-966wr" podStartSLOduration=122.769286181 podStartE2EDuration="2m2.769286181s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:26.767218331 +0000 UTC m=+145.696552544" watchObservedRunningTime="2026-02-18 14:57:26.769286181 +0000 UTC m=+145.698620394" Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.792110 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:26 crc kubenswrapper[4896]: E0218 14:57:26.792332 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:27.292306759 +0000 UTC m=+146.221640982 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.795967 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:26 crc kubenswrapper[4896]: E0218 14:57:26.796433 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:27.296392928 +0000 UTC m=+146.225727141 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.807429 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-z7xdz" event={"ID":"918ef2a1-8828-4b17-b30a-de957f009104","Type":"ContainerStarted","Data":"60816958349fd4a4d9c9a48ff988c30ac051da93dc51108b6d2aa1ad09ed21b0"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.808872 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bjbsw" event={"ID":"b60354fe-e785-4e88-af1c-0d846d9ff59a","Type":"ContainerStarted","Data":"59a1c0311f0c9a8fd3b60107c03b52d4f44ac7d46dbd8f3455070284cf639101"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.811630 4896 generic.go:334] "Generic (PLEG): container finished" podID="23b10824-dc70-459f-b037-6f8f00ae6a04" containerID="8bedfff86a22056b6abaa2da61c74a63ead3dd23769599cf19d11546f27c25dc" exitCode=0 Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.811747 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mcjhx" event={"ID":"23b10824-dc70-459f-b037-6f8f00ae6a04","Type":"ContainerDied","Data":"8bedfff86a22056b6abaa2da61c74a63ead3dd23769599cf19d11546f27c25dc"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.819362 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-4bxkk" event={"ID":"fa6c59b4-6823-47ef-a8a6-52c94e53ef4c","Type":"ContainerStarted","Data":"03056e4484d7c6e4b41f9b1198a08b8235747a35e587e6393de847378b71efe4"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.819712 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-4bxkk" event={"ID":"fa6c59b4-6823-47ef-a8a6-52c94e53ef4c","Type":"ContainerStarted","Data":"0df976b6151ef1e4de2de08574da9369865a47d68a2f9fde896b4dc79284861a"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.824211 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-66zwg" event={"ID":"64e32b45-d067-4151-980d-36ea8f3116e9","Type":"ContainerStarted","Data":"1835f655a11b97ff3e75081891c73d8ad9c3677b9cd7b6eaeb0b8ef72c276ca2"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.825749 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pclrg" event={"ID":"0e9fc2ab-6d3d-441d-8591-3d64ce77c62d","Type":"ContainerStarted","Data":"2f137a6b8c6bffe7f213f1c9727f308ed96678b12315f064b0006f53f838e38a"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.828450 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-rx8f2" event={"ID":"2573759f-2db1-452a-b310-f2a170099ec6","Type":"ContainerStarted","Data":"859c5af9015778ca785ec525adb75054a65248fdaa8d8e0921af17e90bf7c275"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.829518 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jpg48" event={"ID":"8c1a0565-e55b-410e-9b63-12ababdb69eb","Type":"ContainerStarted","Data":"db03e5f5ec1706a678e09d06cb2b734ce634a20759d278d60bf933db35c20307"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.832009 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-znxvs" event={"ID":"8fa6844e-65a5-401e-b329-2a49dc5debe4","Type":"ContainerStarted","Data":"5d9fe13d0b8a29a8606e4db8d39484c6047cdab3b6e1946b2702ceb76b1b2424"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.834262 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-ccl77" event={"ID":"8db0982f-0e1d-493c-8b32-64190372fcb9","Type":"ContainerStarted","Data":"e3c4a9d03bc8b8969a13d5556e3c510019742807f001e8ff78d08aa1da6250c3"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.846419 4896 generic.go:334] "Generic (PLEG): container finished" podID="710fc779-629b-4d37-a820-e08f5289677f" containerID="f6e1ae7630a404c8bbfeeb45559847e6bb12baa77950a21722efd9de44b8018a" exitCode=0 Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.846682 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" event={"ID":"710fc779-629b-4d37-a820-e08f5289677f","Type":"ContainerDied","Data":"f6e1ae7630a404c8bbfeeb45559847e6bb12baa77950a21722efd9de44b8018a"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.851489 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7jgjj" event={"ID":"dcf21347-4c2f-4d66-9137-c24ee63a31de","Type":"ContainerStarted","Data":"356f2de82182d03ce892e2593c61a2cdf4eed2b796b5f5f32527eef70fb2cdd0"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.866511 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bw75d" event={"ID":"3cd5b209-b64c-4ce1-85e2-e00e568b4bbe","Type":"ContainerStarted","Data":"e830c993032cb68e051017b479c14c8d395d636bd79a73283bf05a0a9fd75227"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.871753 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" event={"ID":"76cb1e0e-4321-4450-9fcb-b505a7a1d969","Type":"ContainerStarted","Data":"fce8bf5bd3d671ffcdafbfabccda9c929d8b1b87b96a506cadf5a5c962411ee1"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.872291 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.874405 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v59s8" event={"ID":"b0388f1f-8b9c-46e4-a7f0-1fad187d6d65","Type":"ContainerStarted","Data":"c3c73a000c2312101aca6cfa36c59655891ebdc3dd16a00cf47d7851ae65b837"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.875227 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-27bbf" event={"ID":"31b0687b-5c42-4259-b96d-284c1aa487b4","Type":"ContainerStarted","Data":"6709c0c6bc6eb8fad3c1e2c00d5547edbd703aa0fbbbea4809220b10b020f485"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.875272 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-27bbf" event={"ID":"31b0687b-5c42-4259-b96d-284c1aa487b4","Type":"ContainerStarted","Data":"c8fce6024a05cf318e4dc5dafd64d062e48c552ff0be2303465123c97e71e6b8"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.875772 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-27bbf" Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.876653 4896 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-27bbf container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.876691 4896 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-27bbf" podUID="31b0687b-5c42-4259-b96d-284c1aa487b4" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.877398 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-s95mb" event={"ID":"013d2ce9-60ac-4abf-945d-3340ffa55b72","Type":"ContainerStarted","Data":"868b8d1a63b2ed03e9f7b6408d14c78ce8b4cf77dcb39cb4afbdcb38e382d9bc"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.879540 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j99bk" event={"ID":"5c0e215f-d940-415a-90d3-f5faf3e537de","Type":"ContainerStarted","Data":"6f60f7ed3628973a2510c27903b3c9e439b87317628baa285e862fa1222a154e"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.894789 4896 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-shvc6 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.29:6443/healthz\": dial tcp 10.217.0.29:6443: connect: connection refused" start-of-body= Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.894845 4896 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" podUID="76cb1e0e-4321-4450-9fcb-b505a7a1d969" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.29:6443/healthz\": dial tcp 10.217.0.29:6443: connect: connection refused" Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.898185 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:26 crc kubenswrapper[4896]: E0218 14:57:26.898797 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:27.39877847 +0000 UTC m=+146.328112693 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.903493 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-6v4cn" event={"ID":"036e7602-7608-4033-b18d-d859a7eddebd","Type":"ContainerStarted","Data":"1cfd970f6043535dd4405bb5e7883dcadacd5ad7b759142e905dd350cf7bf3f6"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.904343 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-6v4cn" Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.904998 4896 patch_prober.go:28] interesting pod/downloads-7954f5f757-6v4cn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.905025 4896 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6v4cn" podUID="036e7602-7608-4033-b18d-d859a7eddebd" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.911042 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ftds" event={"ID":"71145266-38d8-495b-83c2-cff3c68e705d","Type":"ContainerStarted","Data":"eb10b58ddfd2af1e93d20a05d0fa8de78150be6882e4f675610bd2095bc3f8bb"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.911074 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ftds" event={"ID":"71145266-38d8-495b-83c2-cff3c68e705d","Type":"ContainerStarted","Data":"5fd14f68af55118cc1c6b3d313fb5c00ca0d1485cc73bd17a7542075ac8c1deb"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.923310 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l8fcf" event={"ID":"7b9d6a43-e957-4d02-9928-020f9f17f7a9","Type":"ContainerStarted","Data":"6055e5db8473adde5b1e41e95b1353c48aa9b9a2efa6a20acd918a75129847df"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.923354 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l8fcf" event={"ID":"7b9d6a43-e957-4d02-9928-020f9f17f7a9","Type":"ContainerStarted","Data":"24e7b47ef018a882ec8f37358a12ecb5feca603d127b25f4bd35bef6a4c64e57"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.923807 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l8fcf" Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.932994 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hv6vt" event={"ID":"86b5fc8f-1b76-43ad-b283-0d46697be7aa","Type":"ContainerStarted","Data":"d0346143cd569e223ae5886f0579a05e37795a700fae558370355c6f1cfde98b"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.938426 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dlghr" event={"ID":"37bf7cb8-e127-458a-ab07-efb2c111ce02","Type":"ContainerStarted","Data":"00c8af96c21f5e41c6ec20cbe8bbff6540ba1603e2d41da18996e24b431f86f3"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.938467 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dlghr" event={"ID":"37bf7cb8-e127-458a-ab07-efb2c111ce02","Type":"ContainerStarted","Data":"a90e81b3901e748142ec9fc03b2d9ea169d5bab2205f6dda6891aa0e07a1d7ad"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.939978 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-rmc7c" event={"ID":"17344065-a988-48bb-aff9-d51ffd76f54b","Type":"ContainerStarted","Data":"246cc2110bb1327de068be3c63c95d6d21f7dd1f0bd38f3fe57370477a24d004"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.940002 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-rmc7c" event={"ID":"17344065-a988-48bb-aff9-d51ffd76f54b","Type":"ContainerStarted","Data":"8cafd9ed1f8b1e64fab031d430dd2ddfef68b68f180fc3ed30b925b48793dc83"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.940980 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-5795c" event={"ID":"9feb8b3f-6359-41f1-8648-c827ce9c5ec4","Type":"ContainerStarted","Data":"ca86224b4f5abb1167af7ee1a7c9824ae4da437aa4eb5e5b4f77ec3e8aecf5b5"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.941983 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523765-frmvz" event={"ID":"dd10c491-df5d-43a1-961d-613e2a90be23","Type":"ContainerStarted","Data":"11146cb2df71c24c702376b35aacefde86b1d76f004e5376a7e387f42f337042"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.944146 4896 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-l8fcf container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" start-of-body= Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.944182 4896 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l8fcf" podUID="7b9d6a43-e957-4d02-9928-020f9f17f7a9" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.944848 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gqb6w" event={"ID":"f8abcf75-9577-4dfe-a252-40aff339ffce","Type":"ContainerStarted","Data":"ff6a0b5ded30ffd201e567430a67b18f842001a1efc247a76897559667881aad"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.974549 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-fhpqt" event={"ID":"6225d3ae-2a4e-479c-aa70-6fdbd2ee488b","Type":"ContainerStarted","Data":"918336ad0174b2748647a075fab305b5b2c79575cb814a52ad63e3b6d29dc523"} Feb 18 14:57:26 crc kubenswrapper[4896]: I0218 14:57:26.974601 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-fhpqt" event={"ID":"6225d3ae-2a4e-479c-aa70-6fdbd2ee488b","Type":"ContainerStarted","Data":"29e358b2a7b9a40e72dfae313975c2b37788f17ac208598111b344cd774f447e"} Feb 18 14:57:27 crc kubenswrapper[4896]: I0218 14:57:27.000319 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:27 crc kubenswrapper[4896]: I0218 14:57:27.002491 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6frkl" event={"ID":"21fb3328-3882-4644-9a94-b7949f48db54","Type":"ContainerStarted","Data":"8ae03f773b1a13146efe0776c38b13b92858dda012aa96b94bbb489715c4bc1c"} Feb 18 14:57:27 crc kubenswrapper[4896]: I0218 14:57:27.002540 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6frkl" event={"ID":"21fb3328-3882-4644-9a94-b7949f48db54","Type":"ContainerStarted","Data":"4a1708d85f2fcfce070ae6a3ecae558ef3b1e140b9d84cf2e1c31e3814c0cfb5"} Feb 18 14:57:27 crc kubenswrapper[4896]: E0218 14:57:27.031552 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:27.531532625 +0000 UTC m=+146.460866838 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:27 crc kubenswrapper[4896]: I0218 14:57:27.066918 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bncs" event={"ID":"4122c286-47e3-460d-bd25-323012847f12","Type":"ContainerStarted","Data":"6be157b59b10da20b2133b7f29f3d3394d96a9f83da466e4b6d863bfa1c0e854"} Feb 18 14:57:27 crc kubenswrapper[4896]: I0218 14:57:27.067462 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bncs" event={"ID":"4122c286-47e3-460d-bd25-323012847f12","Type":"ContainerStarted","Data":"9f5686e74dbe7cd311bdf4eefcdce50a487bb984acfbc94b6ae49dcc1081556e"} Feb 18 14:57:27 crc kubenswrapper[4896]: I0218 14:57:27.085310 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xrhcj" event={"ID":"0f95dce4-6d68-44ac-9dd3-215664bb1d75","Type":"ContainerStarted","Data":"67c9b838de57980c73fed1e514d27d34433e345ed4f732650b1e366e7da115ca"} Feb 18 14:57:27 crc kubenswrapper[4896]: I0218 14:57:27.085859 4896 patch_prober.go:28] interesting pod/console-operator-58897d9998-p7zgl container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Feb 18 14:57:27 crc kubenswrapper[4896]: I0218 14:57:27.085897 4896 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-p7zgl" podUID="a1659f4f-bd6e-4d58-a56b-69b14ab57ed0" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" Feb 18 14:57:27 crc kubenswrapper[4896]: I0218 14:57:27.111804 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:27 crc kubenswrapper[4896]: E0218 14:57:27.112462 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:27.612435854 +0000 UTC m=+146.541770067 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:27 crc kubenswrapper[4896]: I0218 14:57:27.220252 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-zhgdz" podStartSLOduration=123.220219773 podStartE2EDuration="2m3.220219773s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:27.2187396 +0000 UTC m=+146.148073813" watchObservedRunningTime="2026-02-18 14:57:27.220219773 +0000 UTC m=+146.149553986" Feb 18 14:57:27 crc kubenswrapper[4896]: I0218 14:57:27.222967 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:27 crc kubenswrapper[4896]: E0218 14:57:27.225916 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:27.725898778 +0000 UTC m=+146.655233061 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:27 crc kubenswrapper[4896]: I0218 14:57:27.257771 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-n25dz" podStartSLOduration=123.257756103 podStartE2EDuration="2m3.257756103s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:27.257588138 +0000 UTC m=+146.186922351" watchObservedRunningTime="2026-02-18 14:57:27.257756103 +0000 UTC m=+146.187090316" Feb 18 14:57:27 crc kubenswrapper[4896]: I0218 14:57:27.328673 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:27 crc kubenswrapper[4896]: E0218 14:57:27.329081 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:27.829065543 +0000 UTC m=+146.758399756 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:27 crc kubenswrapper[4896]: I0218 14:57:27.340304 4896 patch_prober.go:28] interesting pod/router-default-5444994796-8cq46 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 14:57:27 crc kubenswrapper[4896]: [-]has-synced failed: reason withheld Feb 18 14:57:27 crc kubenswrapper[4896]: [+]process-running ok Feb 18 14:57:27 crc kubenswrapper[4896]: healthz check failed Feb 18 14:57:27 crc kubenswrapper[4896]: I0218 14:57:27.340351 4896 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8cq46" podUID="7a51266e-326e-4b5c-a25a-fb61bf1f3757" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 14:57:27 crc kubenswrapper[4896]: I0218 14:57:27.432061 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:27 crc kubenswrapper[4896]: E0218 14:57:27.433251 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:27.933223987 +0000 UTC m=+146.862558200 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:27 crc kubenswrapper[4896]: I0218 14:57:27.489540 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-fjmq5" podStartSLOduration=123.489521961 podStartE2EDuration="2m3.489521961s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:27.451313172 +0000 UTC m=+146.380647385" watchObservedRunningTime="2026-02-18 14:57:27.489521961 +0000 UTC m=+146.418856174" Feb 18 14:57:27 crc kubenswrapper[4896]: I0218 14:57:27.533057 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:27 crc kubenswrapper[4896]: E0218 14:57:27.533419 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:28.033406366 +0000 UTC m=+146.962740579 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:27 crc kubenswrapper[4896]: I0218 14:57:27.540977 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" podStartSLOduration=123.540960455 podStartE2EDuration="2m3.540960455s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:27.540479311 +0000 UTC m=+146.469813544" watchObservedRunningTime="2026-02-18 14:57:27.540960455 +0000 UTC m=+146.470294668" Feb 18 14:57:27 crc kubenswrapper[4896]: I0218 14:57:27.634029 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:27 crc kubenswrapper[4896]: E0218 14:57:27.634655 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:28.134640455 +0000 UTC m=+147.063974668 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:27 crc kubenswrapper[4896]: I0218 14:57:27.685282 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-27bbf" podStartSLOduration=123.685260114 podStartE2EDuration="2m3.685260114s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:27.684963866 +0000 UTC m=+146.614298079" watchObservedRunningTime="2026-02-18 14:57:27.685260114 +0000 UTC m=+146.614594337" Feb 18 14:57:27 crc kubenswrapper[4896]: I0218 14:57:27.722264 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bncs" podStartSLOduration=123.722247408 podStartE2EDuration="2m3.722247408s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:27.719989843 +0000 UTC m=+146.649324056" watchObservedRunningTime="2026-02-18 14:57:27.722247408 +0000 UTC m=+146.651581621" Feb 18 14:57:27 crc kubenswrapper[4896]: I0218 14:57:27.734912 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:27 crc kubenswrapper[4896]: E0218 14:57:27.735065 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:28.23504185 +0000 UTC m=+147.164376063 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:27 crc kubenswrapper[4896]: I0218 14:57:27.735361 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:27 crc kubenswrapper[4896]: E0218 14:57:27.735658 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:28.235650237 +0000 UTC m=+147.164984450 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:27 crc kubenswrapper[4896]: I0218 14:57:27.770778 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-dlghr" podStartSLOduration=123.770760917 podStartE2EDuration="2m3.770760917s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:27.769368266 +0000 UTC m=+146.698702479" watchObservedRunningTime="2026-02-18 14:57:27.770760917 +0000 UTC m=+146.700095130" Feb 18 14:57:27 crc kubenswrapper[4896]: I0218 14:57:27.804969 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l8fcf" podStartSLOduration=123.804943939 podStartE2EDuration="2m3.804943939s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:27.803508127 +0000 UTC m=+146.732842340" watchObservedRunningTime="2026-02-18 14:57:27.804943939 +0000 UTC m=+146.734278152" Feb 18 14:57:27 crc kubenswrapper[4896]: I0218 14:57:27.835939 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:27 crc kubenswrapper[4896]: E0218 14:57:27.836399 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:28.336380212 +0000 UTC m=+147.265714425 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:27 crc kubenswrapper[4896]: I0218 14:57:27.845451 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-znxvs" podStartSLOduration=5.845436205 podStartE2EDuration="5.845436205s" podCreationTimestamp="2026-02-18 14:57:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:27.843273292 +0000 UTC m=+146.772607505" watchObservedRunningTime="2026-02-18 14:57:27.845436205 +0000 UTC m=+146.774770418" Feb 18 14:57:27 crc kubenswrapper[4896]: I0218 14:57:27.929718 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bw75d" podStartSLOduration=123.929697251 podStartE2EDuration="2m3.929697251s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:27.928447745 +0000 UTC m=+146.857781958" watchObservedRunningTime="2026-02-18 14:57:27.929697251 +0000 UTC m=+146.859031464" Feb 18 14:57:27 crc kubenswrapper[4896]: I0218 14:57:27.930996 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-4bxkk" podStartSLOduration=123.930988449 podStartE2EDuration="2m3.930988449s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:27.896900959 +0000 UTC m=+146.826235172" watchObservedRunningTime="2026-02-18 14:57:27.930988449 +0000 UTC m=+146.860322662" Feb 18 14:57:27 crc kubenswrapper[4896]: I0218 14:57:27.939722 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:27 crc kubenswrapper[4896]: E0218 14:57:27.940018 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:28.44000761 +0000 UTC m=+147.369341823 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:27 crc kubenswrapper[4896]: I0218 14:57:27.973257 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-gqb6w" podStartSLOduration=123.973223855 podStartE2EDuration="2m3.973223855s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:27.970227748 +0000 UTC m=+146.899561961" watchObservedRunningTime="2026-02-18 14:57:27.973223855 +0000 UTC m=+146.902558068" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.040555 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:28 crc kubenswrapper[4896]: E0218 14:57:28.040978 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:28.540962441 +0000 UTC m=+147.470296654 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.096735 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6frkl" podStartSLOduration=124.09672101 podStartE2EDuration="2m4.09672101s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:28.045850793 +0000 UTC m=+146.975185006" watchObservedRunningTime="2026-02-18 14:57:28.09672101 +0000 UTC m=+147.026055223" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.102759 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-rx8f2" event={"ID":"2573759f-2db1-452a-b310-f2a170099ec6","Type":"ContainerStarted","Data":"0a551155051d5a2f807d1ed1f96e999c94cf09d3b8117ad5ebe09d184f86f171"} Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.104603 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jpg48" event={"ID":"8c1a0565-e55b-410e-9b63-12ababdb69eb","Type":"ContainerStarted","Data":"6fe98f7fbbe4f7553f7aa32ab2145e5875dfa511f9f437a4c8edeab3268b57a9"} Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.105955 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-z7xdz" event={"ID":"918ef2a1-8828-4b17-b30a-de957f009104","Type":"ContainerStarted","Data":"fb1260908a8952a8417f33f15f2281e2b6d4a2ddf149874e7320dc648cf68692"} Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.109150 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v59s8" event={"ID":"b0388f1f-8b9c-46e4-a7f0-1fad187d6d65","Type":"ContainerStarted","Data":"66659f8b4052ba53ce7aa453f2c13e4dd96419b32b4172ef2bf03256036b866d"} Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.109188 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v59s8" event={"ID":"b0388f1f-8b9c-46e4-a7f0-1fad187d6d65","Type":"ContainerStarted","Data":"1291fe7d3af6aad02452bc6ca63e9a53ddee22b10846c798dd7160c9ea1bae8c"} Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.113217 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mcjhx" event={"ID":"23b10824-dc70-459f-b037-6f8f00ae6a04","Type":"ContainerStarted","Data":"6ec3d5e40477be454557c2c53eca50f832e29564dda986f3364bed0035be3cc3"} Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.113797 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mcjhx" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.125512 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" event={"ID":"710fc779-629b-4d37-a820-e08f5289677f","Type":"ContainerStarted","Data":"2af6b008d2406db0911b91f2977b19e7b956909185272e1f967693b8e124f313"} Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.128969 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2ftds" podStartSLOduration=124.128950906 podStartE2EDuration="2m4.128950906s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:28.096560716 +0000 UTC m=+147.025894929" watchObservedRunningTime="2026-02-18 14:57:28.128950906 +0000 UTC m=+147.058285119" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.132730 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bjbsw" event={"ID":"b60354fe-e785-4e88-af1c-0d846d9ff59a","Type":"ContainerStarted","Data":"7aa20320c8c3df9e5a8182cb75df2211eb63dc7df15fd12f343267cdd010fe31"} Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.137948 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j99bk" event={"ID":"5c0e215f-d940-415a-90d3-f5faf3e537de","Type":"ContainerStarted","Data":"1d309cccd362c3fc65ce63e7b0329dd6c26b441cbd907b161fba4fe492255305"} Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.138625 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j99bk" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.141639 4896 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-j99bk container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.42:8443/healthz\": dial tcp 10.217.0.42:8443: connect: connection refused" start-of-body= Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.141686 4896 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j99bk" podUID="5c0e215f-d940-415a-90d3-f5faf3e537de" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.42:8443/healthz\": dial tcp 10.217.0.42:8443: connect: connection refused" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.141823 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:28 crc kubenswrapper[4896]: E0218 14:57:28.142035 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:28.642015645 +0000 UTC m=+147.571349938 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.143621 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-ccl77" event={"ID":"8db0982f-0e1d-493c-8b32-64190372fcb9","Type":"ContainerStarted","Data":"61f5a280b6fe5ff8061a25fd9f44c44e28257e074f3347a2b8cb64e509975089"} Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.146962 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523765-frmvz" event={"ID":"dd10c491-df5d-43a1-961d-613e2a90be23","Type":"ContainerStarted","Data":"cd87c33df1401c30da4665cb409b6146edfd54db31a21e42e7551f0db40a8e40"} Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.148985 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-rmc7c" event={"ID":"17344065-a988-48bb-aff9-d51ffd76f54b","Type":"ContainerStarted","Data":"40bddee6ebb9c80f7814077ac51301753034e8f811af8355caa7d8bcf616b1f4"} Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.163194 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hv6vt" event={"ID":"86b5fc8f-1b76-43ad-b283-0d46697be7aa","Type":"ContainerStarted","Data":"9f086050a76b58c82e32224a5679dd5f8fb2dbbe5b307585eef2192cf9b6e089"} Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.163254 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hv6vt" event={"ID":"86b5fc8f-1b76-43ad-b283-0d46697be7aa","Type":"ContainerStarted","Data":"b90398c0e969439f7ce28e712adde20f3d9765d79b50cd2e41b03afdcfd7fb28"} Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.163867 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hv6vt" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.168514 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-5795c" event={"ID":"9feb8b3f-6359-41f1-8648-c827ce9c5ec4","Type":"ContainerStarted","Data":"add423d70e2c7c3ed09967a50a80a55488563710ba1339576fc4ae95162e72f2"} Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.168966 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-5795c" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.170692 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xrhcj" event={"ID":"0f95dce4-6d68-44ac-9dd3-215664bb1d75","Type":"ContainerStarted","Data":"95e1564d67b45208daa6c2d917ef9ddc045a644295fdcc1557bdf6c9fdbea897"} Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.171571 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-xrhcj" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.175432 4896 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-xrhcj container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.175517 4896 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-xrhcj" podUID="0f95dce4-6d68-44ac-9dd3-215664bb1d75" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.176503 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-66zwg" event={"ID":"64e32b45-d067-4151-980d-36ea8f3116e9","Type":"ContainerStarted","Data":"dc341a8894283817e7a372dbd1ee962284ec5bc6e1d5658cd99063fccfb02ef7"} Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.176771 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-66zwg" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.183772 4896 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-66zwg container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.183838 4896 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-66zwg" podUID="64e32b45-d067-4151-980d-36ea8f3116e9" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.184408 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-s95mb" podStartSLOduration=7.184387395 podStartE2EDuration="7.184387395s" podCreationTimestamp="2026-02-18 14:57:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:28.135569438 +0000 UTC m=+147.064903651" watchObservedRunningTime="2026-02-18 14:57:28.184387395 +0000 UTC m=+147.113721608" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.196790 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pclrg" event={"ID":"0e9fc2ab-6d3d-441d-8591-3d64ce77c62d","Type":"ContainerStarted","Data":"22f3ae2a5a8a950e521af0fb1471ddd9889407a1f135101aeea05804306bbc58"} Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.196825 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pclrg" event={"ID":"0e9fc2ab-6d3d-441d-8591-3d64ce77c62d","Type":"ContainerStarted","Data":"14fc34f63296703eb08dda6fefb17bd1673ac13bc0f5d1734b8fb7ded5805073"} Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.196838 4896 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-27bbf container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.196888 4896 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-27bbf" podUID="31b0687b-5c42-4259-b96d-284c1aa487b4" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.197262 4896 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-shvc6 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.29:6443/healthz\": dial tcp 10.217.0.29:6443: connect: connection refused" start-of-body= Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.197293 4896 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" podUID="76cb1e0e-4321-4450-9fcb-b505a7a1d969" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.29:6443/healthz\": dial tcp 10.217.0.29:6443: connect: connection refused" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.197392 4896 patch_prober.go:28] interesting pod/downloads-7954f5f757-6v4cn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.197416 4896 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6v4cn" podUID="036e7602-7608-4033-b18d-d859a7eddebd" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.197673 4896 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-l8fcf container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" start-of-body= Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.197712 4896 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l8fcf" podUID="7b9d6a43-e957-4d02-9928-020f9f17f7a9" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.32:5443/healthz\": dial tcp 10.217.0.32:5443: connect: connection refused" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.212183 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-fhpqt" podStartSLOduration=124.212168772 podStartE2EDuration="2m4.212168772s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:28.210913136 +0000 UTC m=+147.140247369" watchObservedRunningTime="2026-02-18 14:57:28.212168772 +0000 UTC m=+147.141502985" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.243119 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:28 crc kubenswrapper[4896]: E0218 14:57:28.245447 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:28.745422708 +0000 UTC m=+147.674757001 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.254659 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-6v4cn" podStartSLOduration=124.254642725 podStartE2EDuration="2m4.254642725s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:28.254294365 +0000 UTC m=+147.183628578" watchObservedRunningTime="2026-02-18 14:57:28.254642725 +0000 UTC m=+147.183976938" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.323249 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bjbsw" podStartSLOduration=124.323215556 podStartE2EDuration="2m4.323215556s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:28.323030891 +0000 UTC m=+147.252365114" watchObservedRunningTime="2026-02-18 14:57:28.323215556 +0000 UTC m=+147.252549769" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.323972 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" podStartSLOduration=124.323965848 podStartE2EDuration="2m4.323965848s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:28.292282998 +0000 UTC m=+147.221617211" watchObservedRunningTime="2026-02-18 14:57:28.323965848 +0000 UTC m=+147.253300061" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.343671 4896 patch_prober.go:28] interesting pod/router-default-5444994796-8cq46 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 14:57:28 crc kubenswrapper[4896]: [-]has-synced failed: reason withheld Feb 18 14:57:28 crc kubenswrapper[4896]: [+]process-running ok Feb 18 14:57:28 crc kubenswrapper[4896]: healthz check failed Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.343788 4896 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8cq46" podUID="7a51266e-326e-4b5c-a25a-fb61bf1f3757" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.360107 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:28 crc kubenswrapper[4896]: E0218 14:57:28.362836 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:28.862819496 +0000 UTC m=+147.792153709 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.374594 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-ccl77" podStartSLOduration=124.374581017 podStartE2EDuration="2m4.374581017s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:28.374058472 +0000 UTC m=+147.303392685" watchObservedRunningTime="2026-02-18 14:57:28.374581017 +0000 UTC m=+147.303915230" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.412267 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mcjhx" podStartSLOduration=124.41222242 podStartE2EDuration="2m4.41222242s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:28.409946784 +0000 UTC m=+147.339281007" watchObservedRunningTime="2026-02-18 14:57:28.41222242 +0000 UTC m=+147.341556633" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.453201 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hv6vt" podStartSLOduration=124.453182309 podStartE2EDuration="2m4.453182309s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:28.45079742 +0000 UTC m=+147.380131633" watchObservedRunningTime="2026-02-18 14:57:28.453182309 +0000 UTC m=+147.382516522" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.461698 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:28 crc kubenswrapper[4896]: E0218 14:57:28.461985 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:28.961923543 +0000 UTC m=+147.891257756 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.462345 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:28 crc kubenswrapper[4896]: E0218 14:57:28.462651 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:28.962638824 +0000 UTC m=+147.891973037 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.484008 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-jpg48" podStartSLOduration=124.483984574 podStartE2EDuration="2m4.483984574s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:28.482603213 +0000 UTC m=+147.411937426" watchObservedRunningTime="2026-02-18 14:57:28.483984574 +0000 UTC m=+147.413318787" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.529575 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.529640 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.530313 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29523765-frmvz" podStartSLOduration=124.530298268 podStartE2EDuration="2m4.530298268s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:28.527924269 +0000 UTC m=+147.457258482" watchObservedRunningTime="2026-02-18 14:57:28.530298268 +0000 UTC m=+147.459632481" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.548685 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.565196 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.565618 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-z7xdz" podStartSLOduration=124.565598033 podStartE2EDuration="2m4.565598033s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:28.563404169 +0000 UTC m=+147.492738392" watchObservedRunningTime="2026-02-18 14:57:28.565598033 +0000 UTC m=+147.494932256" Feb 18 14:57:28 crc kubenswrapper[4896]: E0218 14:57:28.565905 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:29.065884721 +0000 UTC m=+147.995218954 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.616113 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-66zwg" podStartSLOduration=124.616098189 podStartE2EDuration="2m4.616098189s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:28.61543621 +0000 UTC m=+147.544770423" watchObservedRunningTime="2026-02-18 14:57:28.616098189 +0000 UTC m=+147.545432392" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.652715 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j99bk" podStartSLOduration=124.652693702 podStartE2EDuration="2m4.652693702s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:28.650090216 +0000 UTC m=+147.579424429" watchObservedRunningTime="2026-02-18 14:57:28.652693702 +0000 UTC m=+147.582027925" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.672545 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:28 crc kubenswrapper[4896]: E0218 14:57:28.672924 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:29.172910259 +0000 UTC m=+148.102244472 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.692893 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-v59s8" podStartSLOduration=124.692874018 podStartE2EDuration="2m4.692874018s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:28.68811637 +0000 UTC m=+147.617450603" watchObservedRunningTime="2026-02-18 14:57:28.692874018 +0000 UTC m=+147.622208231" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.773316 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:28 crc kubenswrapper[4896]: E0218 14:57:28.773756 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:29.273734266 +0000 UTC m=+148.203068479 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.781614 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-rx8f2" podStartSLOduration=124.781594584 podStartE2EDuration="2m4.781594584s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:28.744409515 +0000 UTC m=+147.673743738" watchObservedRunningTime="2026-02-18 14:57:28.781594584 +0000 UTC m=+147.710928797" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.783581 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-rmc7c" podStartSLOduration=124.783572822 podStartE2EDuration="2m4.783572822s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:28.780842952 +0000 UTC m=+147.710177165" watchObservedRunningTime="2026-02-18 14:57:28.783572822 +0000 UTC m=+147.712907035" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.817175 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-xrhcj" podStartSLOduration=124.817160837 podStartE2EDuration="2m4.817160837s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:28.816516118 +0000 UTC m=+147.745850351" watchObservedRunningTime="2026-02-18 14:57:28.817160837 +0000 UTC m=+147.746495050" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.850985 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pclrg" podStartSLOduration=124.850963118 podStartE2EDuration="2m4.850963118s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:28.849040422 +0000 UTC m=+147.778374645" watchObservedRunningTime="2026-02-18 14:57:28.850963118 +0000 UTC m=+147.780297341" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.875565 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:28 crc kubenswrapper[4896]: E0218 14:57:28.876008 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:29.375989065 +0000 UTC m=+148.305323328 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.901449 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-5795c" podStartSLOduration=7.901422933 podStartE2EDuration="7.901422933s" podCreationTimestamp="2026-02-18 14:57:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:28.898491188 +0000 UTC m=+147.827825401" watchObservedRunningTime="2026-02-18 14:57:28.901422933 +0000 UTC m=+147.830757146" Feb 18 14:57:28 crc kubenswrapper[4896]: I0218 14:57:28.976989 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:28 crc kubenswrapper[4896]: E0218 14:57:28.977324 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:29.477297926 +0000 UTC m=+148.406632139 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:29 crc kubenswrapper[4896]: I0218 14:57:29.078271 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:29 crc kubenswrapper[4896]: E0218 14:57:29.078562 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:29.578550566 +0000 UTC m=+148.507884779 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:29 crc kubenswrapper[4896]: I0218 14:57:29.180001 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:29 crc kubenswrapper[4896]: E0218 14:57:29.180434 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:29.680416182 +0000 UTC m=+148.609750415 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:29 crc kubenswrapper[4896]: I0218 14:57:29.202206 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" event={"ID":"710fc779-629b-4d37-a820-e08f5289677f","Type":"ContainerStarted","Data":"30d1bd20b5c81074abba90d093daf3b989ad610669eb4f4c506c575ae90f9324"} Feb 18 14:57:29 crc kubenswrapper[4896]: I0218 14:57:29.205098 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7jgjj" event={"ID":"dcf21347-4c2f-4d66-9137-c24ee63a31de","Type":"ContainerStarted","Data":"5931950a194b2f0b2e697d75f776449fa5b2fd3ebae115f209e4ef807073816f"} Feb 18 14:57:29 crc kubenswrapper[4896]: I0218 14:57:29.206062 4896 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-27bbf container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Feb 18 14:57:29 crc kubenswrapper[4896]: I0218 14:57:29.206069 4896 patch_prober.go:28] interesting pod/downloads-7954f5f757-6v4cn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Feb 18 14:57:29 crc kubenswrapper[4896]: I0218 14:57:29.206104 4896 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-27bbf" podUID="31b0687b-5c42-4259-b96d-284c1aa487b4" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Feb 18 14:57:29 crc kubenswrapper[4896]: I0218 14:57:29.206122 4896 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6v4cn" podUID="036e7602-7608-4033-b18d-d859a7eddebd" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Feb 18 14:57:29 crc kubenswrapper[4896]: I0218 14:57:29.206063 4896 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-xrhcj container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Feb 18 14:57:29 crc kubenswrapper[4896]: I0218 14:57:29.206171 4896 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-xrhcj" podUID="0f95dce4-6d68-44ac-9dd3-215664bb1d75" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" Feb 18 14:57:29 crc kubenswrapper[4896]: I0218 14:57:29.220166 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-b8czr" Feb 18 14:57:29 crc kubenswrapper[4896]: I0218 14:57:29.223284 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-66zwg" Feb 18 14:57:29 crc kubenswrapper[4896]: I0218 14:57:29.225538 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" podStartSLOduration=125.225521642 podStartE2EDuration="2m5.225521642s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:29.224608925 +0000 UTC m=+148.153943148" watchObservedRunningTime="2026-02-18 14:57:29.225521642 +0000 UTC m=+148.154855855" Feb 18 14:57:29 crc kubenswrapper[4896]: I0218 14:57:29.235918 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j99bk" Feb 18 14:57:29 crc kubenswrapper[4896]: I0218 14:57:29.282455 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:29 crc kubenswrapper[4896]: E0218 14:57:29.285912 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:29.785896284 +0000 UTC m=+148.715230497 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:29 crc kubenswrapper[4896]: I0218 14:57:29.342565 4896 patch_prober.go:28] interesting pod/router-default-5444994796-8cq46 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 14:57:29 crc kubenswrapper[4896]: [-]has-synced failed: reason withheld Feb 18 14:57:29 crc kubenswrapper[4896]: [+]process-running ok Feb 18 14:57:29 crc kubenswrapper[4896]: healthz check failed Feb 18 14:57:29 crc kubenswrapper[4896]: I0218 14:57:29.342874 4896 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8cq46" podUID="7a51266e-326e-4b5c-a25a-fb61bf1f3757" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 14:57:29 crc kubenswrapper[4896]: I0218 14:57:29.383644 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:29 crc kubenswrapper[4896]: E0218 14:57:29.384269 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:29.884215779 +0000 UTC m=+148.813550002 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:29 crc kubenswrapper[4896]: I0218 14:57:29.485375 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:29 crc kubenswrapper[4896]: E0218 14:57:29.485742 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:29.985729346 +0000 UTC m=+148.915063559 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:29 crc kubenswrapper[4896]: I0218 14:57:29.554860 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:29 crc kubenswrapper[4896]: I0218 14:57:29.554909 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:29 crc kubenswrapper[4896]: I0218 14:57:29.556193 4896 patch_prober.go:28] interesting pod/apiserver-76f77b778f-tgwt9 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="Get \"https://10.217.0.12:8443/livez\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Feb 18 14:57:29 crc kubenswrapper[4896]: I0218 14:57:29.556290 4896 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" podUID="710fc779-629b-4d37-a820-e08f5289677f" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.12:8443/livez\": dial tcp 10.217.0.12:8443: connect: connection refused" Feb 18 14:57:29 crc kubenswrapper[4896]: I0218 14:57:29.586427 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:29 crc kubenswrapper[4896]: E0218 14:57:29.586710 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:30.086669857 +0000 UTC m=+149.016004080 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:29 crc kubenswrapper[4896]: I0218 14:57:29.586801 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:29 crc kubenswrapper[4896]: E0218 14:57:29.587258 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:30.087247754 +0000 UTC m=+149.016582007 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:29 crc kubenswrapper[4896]: I0218 14:57:29.597967 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-l8fcf" Feb 18 14:57:29 crc kubenswrapper[4896]: I0218 14:57:29.687555 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:29 crc kubenswrapper[4896]: E0218 14:57:29.687757 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:30.187731081 +0000 UTC m=+149.117065294 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:29 crc kubenswrapper[4896]: I0218 14:57:29.687819 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:29 crc kubenswrapper[4896]: E0218 14:57:29.688155 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:30.188146773 +0000 UTC m=+149.117481066 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:29 crc kubenswrapper[4896]: I0218 14:57:29.789426 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:29 crc kubenswrapper[4896]: E0218 14:57:29.789618 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:30.289591358 +0000 UTC m=+149.218925571 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:29 crc kubenswrapper[4896]: I0218 14:57:29.789686 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:29 crc kubenswrapper[4896]: E0218 14:57:29.790018 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:30.29000859 +0000 UTC m=+149.219342843 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:29 crc kubenswrapper[4896]: I0218 14:57:29.891486 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:29 crc kubenswrapper[4896]: E0218 14:57:29.891662 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:30.391631661 +0000 UTC m=+149.320965874 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:29 crc kubenswrapper[4896]: I0218 14:57:29.892201 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:29 crc kubenswrapper[4896]: E0218 14:57:29.892704 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:30.392688312 +0000 UTC m=+149.322022525 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:29 crc kubenswrapper[4896]: I0218 14:57:29.993708 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:29 crc kubenswrapper[4896]: E0218 14:57:29.993913 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:30.49388326 +0000 UTC m=+149.423217473 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.095820 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:30 crc kubenswrapper[4896]: E0218 14:57:30.096206 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:30.59618691 +0000 UTC m=+149.525521113 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.196881 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:30 crc kubenswrapper[4896]: E0218 14:57:30.197078 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:30.697050218 +0000 UTC m=+149.626384431 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.197360 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:30 crc kubenswrapper[4896]: E0218 14:57:30.197720 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:30.697704667 +0000 UTC m=+149.627038940 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.212166 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7jgjj" event={"ID":"dcf21347-4c2f-4d66-9137-c24ee63a31de","Type":"ContainerStarted","Data":"72f65005d8c5e8033eb339a29cf79d9f9263ff106004e34cb1f7045f03b482c1"} Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.212227 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7jgjj" event={"ID":"dcf21347-4c2f-4d66-9137-c24ee63a31de","Type":"ContainerStarted","Data":"01276b39fa5298bd5721adff7abba85e0b8c6a463cbd4e3249c8bba86391055a"} Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.221786 4896 generic.go:334] "Generic (PLEG): container finished" podID="dd10c491-df5d-43a1-961d-613e2a90be23" containerID="cd87c33df1401c30da4665cb409b6146edfd54db31a21e42e7551f0db40a8e40" exitCode=0 Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.221969 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523765-frmvz" event={"ID":"dd10c491-df5d-43a1-961d-613e2a90be23","Type":"ContainerDied","Data":"cd87c33df1401c30da4665cb409b6146edfd54db31a21e42e7551f0db40a8e40"} Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.288577 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-xrhcj" Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.298037 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:30 crc kubenswrapper[4896]: E0218 14:57:30.298154 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:30.798135543 +0000 UTC m=+149.727469756 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.299531 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:30 crc kubenswrapper[4896]: E0218 14:57:30.299899 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:30.799880834 +0000 UTC m=+149.729215047 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.339976 4896 patch_prober.go:28] interesting pod/router-default-5444994796-8cq46 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 14:57:30 crc kubenswrapper[4896]: [-]has-synced failed: reason withheld Feb 18 14:57:30 crc kubenswrapper[4896]: [+]process-running ok Feb 18 14:57:30 crc kubenswrapper[4896]: healthz check failed Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.340340 4896 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8cq46" podUID="7a51266e-326e-4b5c-a25a-fb61bf1f3757" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.400708 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:30 crc kubenswrapper[4896]: E0218 14:57:30.400894 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:30.900864526 +0000 UTC m=+149.830198739 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.402410 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:30 crc kubenswrapper[4896]: E0218 14:57:30.402652 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:30.902636297 +0000 UTC m=+149.831970510 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.458899 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mcjhx" Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.503215 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:30 crc kubenswrapper[4896]: E0218 14:57:30.503626 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:31.003610809 +0000 UTC m=+149.932945012 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.604542 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:30 crc kubenswrapper[4896]: E0218 14:57:30.604844 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:31.104833667 +0000 UTC m=+150.034167880 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.706057 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:30 crc kubenswrapper[4896]: E0218 14:57:30.706579 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:31.206559161 +0000 UTC m=+150.135893374 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.807965 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.808016 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.808062 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.808084 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.808109 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:57:30 crc kubenswrapper[4896]: E0218 14:57:30.808736 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:31.308724397 +0000 UTC m=+150.238058610 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.811633 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.815642 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.820878 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.828199 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.855313 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-46f87"] Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.856149 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-46f87" Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.858398 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.873760 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-46f87"] Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.909909 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:30 crc kubenswrapper[4896]: E0218 14:57:30.910336 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:31.410320117 +0000 UTC m=+150.339654330 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.943526 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.951569 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 18 14:57:30 crc kubenswrapper[4896]: I0218 14:57:30.962094 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.011301 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gv76m\" (UniqueName: \"kubernetes.io/projected/aa56560c-aab8-481c-9e3e-ac97ed9b69a5-kube-api-access-gv76m\") pod \"certified-operators-46f87\" (UID: \"aa56560c-aab8-481c-9e3e-ac97ed9b69a5\") " pod="openshift-marketplace/certified-operators-46f87" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.011590 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa56560c-aab8-481c-9e3e-ac97ed9b69a5-utilities\") pod \"certified-operators-46f87\" (UID: \"aa56560c-aab8-481c-9e3e-ac97ed9b69a5\") " pod="openshift-marketplace/certified-operators-46f87" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.011633 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.011656 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa56560c-aab8-481c-9e3e-ac97ed9b69a5-catalog-content\") pod \"certified-operators-46f87\" (UID: \"aa56560c-aab8-481c-9e3e-ac97ed9b69a5\") " pod="openshift-marketplace/certified-operators-46f87" Feb 18 14:57:31 crc kubenswrapper[4896]: E0218 14:57:31.011948 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-18 14:57:31.511935787 +0000 UTC m=+150.441270000 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-qrwnx" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.046802 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cp8c2"] Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.062833 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cp8c2" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.065293 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.084656 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cp8c2"] Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.114754 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.114960 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gv76m\" (UniqueName: \"kubernetes.io/projected/aa56560c-aab8-481c-9e3e-ac97ed9b69a5-kube-api-access-gv76m\") pod \"certified-operators-46f87\" (UID: \"aa56560c-aab8-481c-9e3e-ac97ed9b69a5\") " pod="openshift-marketplace/certified-operators-46f87" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.115008 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa56560c-aab8-481c-9e3e-ac97ed9b69a5-utilities\") pod \"certified-operators-46f87\" (UID: \"aa56560c-aab8-481c-9e3e-ac97ed9b69a5\") " pod="openshift-marketplace/certified-operators-46f87" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.115054 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa56560c-aab8-481c-9e3e-ac97ed9b69a5-catalog-content\") pod \"certified-operators-46f87\" (UID: \"aa56560c-aab8-481c-9e3e-ac97ed9b69a5\") " pod="openshift-marketplace/certified-operators-46f87" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.115461 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa56560c-aab8-481c-9e3e-ac97ed9b69a5-catalog-content\") pod \"certified-operators-46f87\" (UID: \"aa56560c-aab8-481c-9e3e-ac97ed9b69a5\") " pod="openshift-marketplace/certified-operators-46f87" Feb 18 14:57:31 crc kubenswrapper[4896]: E0218 14:57:31.115533 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-18 14:57:31.615517884 +0000 UTC m=+150.544852097 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.115928 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa56560c-aab8-481c-9e3e-ac97ed9b69a5-utilities\") pod \"certified-operators-46f87\" (UID: \"aa56560c-aab8-481c-9e3e-ac97ed9b69a5\") " pod="openshift-marketplace/certified-operators-46f87" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.130642 4896 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.131207 4896 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-02-18T14:57:31.130668394Z","Handler":null,"Name":""} Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.146515 4896 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.146558 4896 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.161899 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gv76m\" (UniqueName: \"kubernetes.io/projected/aa56560c-aab8-481c-9e3e-ac97ed9b69a5-kube-api-access-gv76m\") pod \"certified-operators-46f87\" (UID: \"aa56560c-aab8-481c-9e3e-ac97ed9b69a5\") " pod="openshift-marketplace/certified-operators-46f87" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.169014 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-46f87" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.216259 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96fbf244-90e4-4ffb-8dab-1ad893364032-utilities\") pod \"community-operators-cp8c2\" (UID: \"96fbf244-90e4-4ffb-8dab-1ad893364032\") " pod="openshift-marketplace/community-operators-cp8c2" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.216298 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96fbf244-90e4-4ffb-8dab-1ad893364032-catalog-content\") pod \"community-operators-cp8c2\" (UID: \"96fbf244-90e4-4ffb-8dab-1ad893364032\") " pod="openshift-marketplace/community-operators-cp8c2" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.216342 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsqjj\" (UniqueName: \"kubernetes.io/projected/96fbf244-90e4-4ffb-8dab-1ad893364032-kube-api-access-hsqjj\") pod \"community-operators-cp8c2\" (UID: \"96fbf244-90e4-4ffb-8dab-1ad893364032\") " pod="openshift-marketplace/community-operators-cp8c2" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.216406 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.220907 4896 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.221183 4896 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.237600 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-7jgjj" event={"ID":"dcf21347-4c2f-4d66-9137-c24ee63a31de","Type":"ContainerStarted","Data":"317f08684fa0b6cc37ce14785a8153f55bbc4aa0a06f64a03d6d28aa61f30856"} Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.250129 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7s2vj"] Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.251346 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7s2vj" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.278202 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7s2vj"] Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.280896 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-7jgjj" podStartSLOduration=10.280874545 podStartE2EDuration="10.280874545s" podCreationTimestamp="2026-02-18 14:57:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:31.270458462 +0000 UTC m=+150.199792675" watchObservedRunningTime="2026-02-18 14:57:31.280874545 +0000 UTC m=+150.210208768" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.317815 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96fbf244-90e4-4ffb-8dab-1ad893364032-utilities\") pod \"community-operators-cp8c2\" (UID: \"96fbf244-90e4-4ffb-8dab-1ad893364032\") " pod="openshift-marketplace/community-operators-cp8c2" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.317864 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96fbf244-90e4-4ffb-8dab-1ad893364032-catalog-content\") pod \"community-operators-cp8c2\" (UID: \"96fbf244-90e4-4ffb-8dab-1ad893364032\") " pod="openshift-marketplace/community-operators-cp8c2" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.317895 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsqjj\" (UniqueName: \"kubernetes.io/projected/96fbf244-90e4-4ffb-8dab-1ad893364032-kube-api-access-hsqjj\") pod \"community-operators-cp8c2\" (UID: \"96fbf244-90e4-4ffb-8dab-1ad893364032\") " pod="openshift-marketplace/community-operators-cp8c2" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.318652 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96fbf244-90e4-4ffb-8dab-1ad893364032-utilities\") pod \"community-operators-cp8c2\" (UID: \"96fbf244-90e4-4ffb-8dab-1ad893364032\") " pod="openshift-marketplace/community-operators-cp8c2" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.318806 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96fbf244-90e4-4ffb-8dab-1ad893364032-catalog-content\") pod \"community-operators-cp8c2\" (UID: \"96fbf244-90e4-4ffb-8dab-1ad893364032\") " pod="openshift-marketplace/community-operators-cp8c2" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.328062 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-qrwnx\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.340128 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsqjj\" (UniqueName: \"kubernetes.io/projected/96fbf244-90e4-4ffb-8dab-1ad893364032-kube-api-access-hsqjj\") pod \"community-operators-cp8c2\" (UID: \"96fbf244-90e4-4ffb-8dab-1ad893364032\") " pod="openshift-marketplace/community-operators-cp8c2" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.377054 4896 patch_prober.go:28] interesting pod/router-default-5444994796-8cq46 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 14:57:31 crc kubenswrapper[4896]: [-]has-synced failed: reason withheld Feb 18 14:57:31 crc kubenswrapper[4896]: [+]process-running ok Feb 18 14:57:31 crc kubenswrapper[4896]: healthz check failed Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.377108 4896 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8cq46" podUID="7a51266e-326e-4b5c-a25a-fb61bf1f3757" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.389433 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cp8c2" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.418870 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.419306 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdc2c09e-a873-4f65-8a5f-949b4562cb10-utilities\") pod \"certified-operators-7s2vj\" (UID: \"cdc2c09e-a873-4f65-8a5f-949b4562cb10\") " pod="openshift-marketplace/certified-operators-7s2vj" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.419338 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdc2c09e-a873-4f65-8a5f-949b4562cb10-catalog-content\") pod \"certified-operators-7s2vj\" (UID: \"cdc2c09e-a873-4f65-8a5f-949b4562cb10\") " pod="openshift-marketplace/certified-operators-7s2vj" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.419392 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjjt6\" (UniqueName: \"kubernetes.io/projected/cdc2c09e-a873-4f65-8a5f-949b4562cb10-kube-api-access-bjjt6\") pod \"certified-operators-7s2vj\" (UID: \"cdc2c09e-a873-4f65-8a5f-949b4562cb10\") " pod="openshift-marketplace/certified-operators-7s2vj" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.450777 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wzwx5"] Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.451965 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wzwx5" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.462524 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.463410 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.469197 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.471164 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.464980 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wzwx5"] Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.523081 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.535787 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrtfz\" (UniqueName: \"kubernetes.io/projected/f48b5c5e-4ec2-4bca-bd26-02b837c029ea-kube-api-access-jrtfz\") pod \"community-operators-wzwx5\" (UID: \"f48b5c5e-4ec2-4bca-bd26-02b837c029ea\") " pod="openshift-marketplace/community-operators-wzwx5" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.535827 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f48b5c5e-4ec2-4bca-bd26-02b837c029ea-catalog-content\") pod \"community-operators-wzwx5\" (UID: \"f48b5c5e-4ec2-4bca-bd26-02b837c029ea\") " pod="openshift-marketplace/community-operators-wzwx5" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.535856 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdc2c09e-a873-4f65-8a5f-949b4562cb10-utilities\") pod \"certified-operators-7s2vj\" (UID: \"cdc2c09e-a873-4f65-8a5f-949b4562cb10\") " pod="openshift-marketplace/certified-operators-7s2vj" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.535876 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdc2c09e-a873-4f65-8a5f-949b4562cb10-catalog-content\") pod \"certified-operators-7s2vj\" (UID: \"cdc2c09e-a873-4f65-8a5f-949b4562cb10\") " pod="openshift-marketplace/certified-operators-7s2vj" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.535895 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjjt6\" (UniqueName: \"kubernetes.io/projected/cdc2c09e-a873-4f65-8a5f-949b4562cb10-kube-api-access-bjjt6\") pod \"certified-operators-7s2vj\" (UID: \"cdc2c09e-a873-4f65-8a5f-949b4562cb10\") " pod="openshift-marketplace/certified-operators-7s2vj" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.535940 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f48b5c5e-4ec2-4bca-bd26-02b837c029ea-utilities\") pod \"community-operators-wzwx5\" (UID: \"f48b5c5e-4ec2-4bca-bd26-02b837c029ea\") " pod="openshift-marketplace/community-operators-wzwx5" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.536321 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdc2c09e-a873-4f65-8a5f-949b4562cb10-utilities\") pod \"certified-operators-7s2vj\" (UID: \"cdc2c09e-a873-4f65-8a5f-949b4562cb10\") " pod="openshift-marketplace/certified-operators-7s2vj" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.536522 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdc2c09e-a873-4f65-8a5f-949b4562cb10-catalog-content\") pod \"certified-operators-7s2vj\" (UID: \"cdc2c09e-a873-4f65-8a5f-949b4562cb10\") " pod="openshift-marketplace/certified-operators-7s2vj" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.549798 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.586338 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.595332 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjjt6\" (UniqueName: \"kubernetes.io/projected/cdc2c09e-a873-4f65-8a5f-949b4562cb10-kube-api-access-bjjt6\") pod \"certified-operators-7s2vj\" (UID: \"cdc2c09e-a873-4f65-8a5f-949b4562cb10\") " pod="openshift-marketplace/certified-operators-7s2vj" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.638678 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f48b5c5e-4ec2-4bca-bd26-02b837c029ea-utilities\") pod \"community-operators-wzwx5\" (UID: \"f48b5c5e-4ec2-4bca-bd26-02b837c029ea\") " pod="openshift-marketplace/community-operators-wzwx5" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.638739 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrtfz\" (UniqueName: \"kubernetes.io/projected/f48b5c5e-4ec2-4bca-bd26-02b837c029ea-kube-api-access-jrtfz\") pod \"community-operators-wzwx5\" (UID: \"f48b5c5e-4ec2-4bca-bd26-02b837c029ea\") " pod="openshift-marketplace/community-operators-wzwx5" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.638761 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/52b1f432-3929-4d34-aa46-ea75deb524be-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"52b1f432-3929-4d34-aa46-ea75deb524be\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.638784 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f48b5c5e-4ec2-4bca-bd26-02b837c029ea-catalog-content\") pod \"community-operators-wzwx5\" (UID: \"f48b5c5e-4ec2-4bca-bd26-02b837c029ea\") " pod="openshift-marketplace/community-operators-wzwx5" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.638825 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/52b1f432-3929-4d34-aa46-ea75deb524be-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"52b1f432-3929-4d34-aa46-ea75deb524be\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.639263 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f48b5c5e-4ec2-4bca-bd26-02b837c029ea-utilities\") pod \"community-operators-wzwx5\" (UID: \"f48b5c5e-4ec2-4bca-bd26-02b837c029ea\") " pod="openshift-marketplace/community-operators-wzwx5" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.639695 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f48b5c5e-4ec2-4bca-bd26-02b837c029ea-catalog-content\") pod \"community-operators-wzwx5\" (UID: \"f48b5c5e-4ec2-4bca-bd26-02b837c029ea\") " pod="openshift-marketplace/community-operators-wzwx5" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.670994 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrtfz\" (UniqueName: \"kubernetes.io/projected/f48b5c5e-4ec2-4bca-bd26-02b837c029ea-kube-api-access-jrtfz\") pod \"community-operators-wzwx5\" (UID: \"f48b5c5e-4ec2-4bca-bd26-02b837c029ea\") " pod="openshift-marketplace/community-operators-wzwx5" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.683519 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7s2vj" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.739787 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/52b1f432-3929-4d34-aa46-ea75deb524be-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"52b1f432-3929-4d34-aa46-ea75deb524be\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.739863 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/52b1f432-3929-4d34-aa46-ea75deb524be-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"52b1f432-3929-4d34-aa46-ea75deb524be\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.739963 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/52b1f432-3929-4d34-aa46-ea75deb524be-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"52b1f432-3929-4d34-aa46-ea75deb524be\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.782052 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wzwx5" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.783492 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/52b1f432-3929-4d34-aa46-ea75deb524be-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"52b1f432-3929-4d34-aa46-ea75deb524be\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.801040 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 18 14:57:31 crc kubenswrapper[4896]: W0218 14:57:31.910956 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-ae808ce6a935a9d9295b90897cc39fa3112ed7d195dce984e1f65fd6801ff24a WatchSource:0}: Error finding container ae808ce6a935a9d9295b90897cc39fa3112ed7d195dce984e1f65fd6801ff24a: Status 404 returned error can't find the container with id ae808ce6a935a9d9295b90897cc39fa3112ed7d195dce984e1f65fd6801ff24a Feb 18 14:57:31 crc kubenswrapper[4896]: I0218 14:57:31.990705 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Feb 18 14:57:32 crc kubenswrapper[4896]: W0218 14:57:32.011866 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-7facaccfdf7745ac1bfa04404c8f2492e81ba498ff13b5236e135d18126b7f74 WatchSource:0}: Error finding container 7facaccfdf7745ac1bfa04404c8f2492e81ba498ff13b5236e135d18126b7f74: Status 404 returned error can't find the container with id 7facaccfdf7745ac1bfa04404c8f2492e81ba498ff13b5236e135d18126b7f74 Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.021027 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cp8c2"] Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.032554 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523765-frmvz" Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.130150 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-46f87"] Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.152153 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dd10c491-df5d-43a1-961d-613e2a90be23-config-volume\") pod \"dd10c491-df5d-43a1-961d-613e2a90be23\" (UID: \"dd10c491-df5d-43a1-961d-613e2a90be23\") " Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.152263 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tq74b\" (UniqueName: \"kubernetes.io/projected/dd10c491-df5d-43a1-961d-613e2a90be23-kube-api-access-tq74b\") pod \"dd10c491-df5d-43a1-961d-613e2a90be23\" (UID: \"dd10c491-df5d-43a1-961d-613e2a90be23\") " Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.152293 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dd10c491-df5d-43a1-961d-613e2a90be23-secret-volume\") pod \"dd10c491-df5d-43a1-961d-613e2a90be23\" (UID: \"dd10c491-df5d-43a1-961d-613e2a90be23\") " Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.153749 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd10c491-df5d-43a1-961d-613e2a90be23-config-volume" (OuterVolumeSpecName: "config-volume") pod "dd10c491-df5d-43a1-961d-613e2a90be23" (UID: "dd10c491-df5d-43a1-961d-613e2a90be23"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.157551 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd10c491-df5d-43a1-961d-613e2a90be23-kube-api-access-tq74b" (OuterVolumeSpecName: "kube-api-access-tq74b") pod "dd10c491-df5d-43a1-961d-613e2a90be23" (UID: "dd10c491-df5d-43a1-961d-613e2a90be23"). InnerVolumeSpecName "kube-api-access-tq74b". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.169726 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd10c491-df5d-43a1-961d-613e2a90be23-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "dd10c491-df5d-43a1-961d-613e2a90be23" (UID: "dd10c491-df5d-43a1-961d-613e2a90be23"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.259046 4896 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dd10c491-df5d-43a1-961d-613e2a90be23-config-volume\") on node \"crc\" DevicePath \"\"" Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.259416 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tq74b\" (UniqueName: \"kubernetes.io/projected/dd10c491-df5d-43a1-961d-613e2a90be23-kube-api-access-tq74b\") on node \"crc\" DevicePath \"\"" Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.259435 4896 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dd10c491-df5d-43a1-961d-613e2a90be23-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.287356 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-46f87" event={"ID":"aa56560c-aab8-481c-9e3e-ac97ed9b69a5","Type":"ContainerStarted","Data":"6ecce065ca1d1a22894bf23af68f120122311d31897b9422cd286634dfbac336"} Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.294720 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cp8c2" event={"ID":"96fbf244-90e4-4ffb-8dab-1ad893364032","Type":"ContainerStarted","Data":"3f785dfd2aa6592795013d4b2127b50188c6f501e4d2378ab05789f64d713f50"} Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.295515 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"ae808ce6a935a9d9295b90897cc39fa3112ed7d195dce984e1f65fd6801ff24a"} Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.297539 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"5544df53e313870dea6b87a90dbe5d8efb628969e48c501b70ebeedb01e309ce"} Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.297559 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"280ba157e1ce4a1e8466984b4b75d61fb7ed317f7b6e6b2345a544a907a974bd"} Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.297824 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.302967 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"7facaccfdf7745ac1bfa04404c8f2492e81ba498ff13b5236e135d18126b7f74"} Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.305068 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7s2vj"] Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.317593 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523765-frmvz" Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.317617 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523765-frmvz" event={"ID":"dd10c491-df5d-43a1-961d-613e2a90be23","Type":"ContainerDied","Data":"11146cb2df71c24c702376b35aacefde86b1d76f004e5376a7e387f42f337042"} Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.317668 4896 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11146cb2df71c24c702376b35aacefde86b1d76f004e5376a7e387f42f337042" Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.355639 4896 patch_prober.go:28] interesting pod/router-default-5444994796-8cq46 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 14:57:32 crc kubenswrapper[4896]: [-]has-synced failed: reason withheld Feb 18 14:57:32 crc kubenswrapper[4896]: [+]process-running ok Feb 18 14:57:32 crc kubenswrapper[4896]: healthz check failed Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.355730 4896 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8cq46" podUID="7a51266e-326e-4b5c-a25a-fb61bf1f3757" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.359784 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qrwnx"] Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.412580 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wzwx5"] Feb 18 14:57:32 crc kubenswrapper[4896]: W0218 14:57:32.444747 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf48b5c5e_4ec2_4bca_bd26_02b837c029ea.slice/crio-a1c7b8fcec9fe0141386698656debe691d33b9e1f90570f42b0cd7a244cd9fde WatchSource:0}: Error finding container a1c7b8fcec9fe0141386698656debe691d33b9e1f90570f42b0cd7a244cd9fde: Status 404 returned error can't find the container with id a1c7b8fcec9fe0141386698656debe691d33b9e1f90570f42b0cd7a244cd9fde Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.587770 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 18 14:57:32 crc kubenswrapper[4896]: W0218 14:57:32.593787 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod52b1f432_3929_4d34_aa46_ea75deb524be.slice/crio-905383d6259d1d822dd22bfc09532dec5eba585fabb3b6dad652ebcee0bf99be WatchSource:0}: Error finding container 905383d6259d1d822dd22bfc09532dec5eba585fabb3b6dad652ebcee0bf99be: Status 404 returned error can't find the container with id 905383d6259d1d822dd22bfc09532dec5eba585fabb3b6dad652ebcee0bf99be Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.841508 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-cmxh5"] Feb 18 14:57:32 crc kubenswrapper[4896]: E0218 14:57:32.841769 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd10c491-df5d-43a1-961d-613e2a90be23" containerName="collect-profiles" Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.841789 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd10c491-df5d-43a1-961d-613e2a90be23" containerName="collect-profiles" Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.841892 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd10c491-df5d-43a1-961d-613e2a90be23" containerName="collect-profiles" Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.842588 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cmxh5" Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.844977 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.849468 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cmxh5"] Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.968281 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6bf362d-c626-4e5d-8ff4-6e847ef4eb47-utilities\") pod \"redhat-marketplace-cmxh5\" (UID: \"d6bf362d-c626-4e5d-8ff4-6e847ef4eb47\") " pod="openshift-marketplace/redhat-marketplace-cmxh5" Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.968403 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6bf362d-c626-4e5d-8ff4-6e847ef4eb47-catalog-content\") pod \"redhat-marketplace-cmxh5\" (UID: \"d6bf362d-c626-4e5d-8ff4-6e847ef4eb47\") " pod="openshift-marketplace/redhat-marketplace-cmxh5" Feb 18 14:57:32 crc kubenswrapper[4896]: I0218 14:57:32.968443 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vtkk\" (UniqueName: \"kubernetes.io/projected/d6bf362d-c626-4e5d-8ff4-6e847ef4eb47-kube-api-access-8vtkk\") pod \"redhat-marketplace-cmxh5\" (UID: \"d6bf362d-c626-4e5d-8ff4-6e847ef4eb47\") " pod="openshift-marketplace/redhat-marketplace-cmxh5" Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.069453 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6bf362d-c626-4e5d-8ff4-6e847ef4eb47-catalog-content\") pod \"redhat-marketplace-cmxh5\" (UID: \"d6bf362d-c626-4e5d-8ff4-6e847ef4eb47\") " pod="openshift-marketplace/redhat-marketplace-cmxh5" Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.069499 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vtkk\" (UniqueName: \"kubernetes.io/projected/d6bf362d-c626-4e5d-8ff4-6e847ef4eb47-kube-api-access-8vtkk\") pod \"redhat-marketplace-cmxh5\" (UID: \"d6bf362d-c626-4e5d-8ff4-6e847ef4eb47\") " pod="openshift-marketplace/redhat-marketplace-cmxh5" Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.069560 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6bf362d-c626-4e5d-8ff4-6e847ef4eb47-utilities\") pod \"redhat-marketplace-cmxh5\" (UID: \"d6bf362d-c626-4e5d-8ff4-6e847ef4eb47\") " pod="openshift-marketplace/redhat-marketplace-cmxh5" Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.070081 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6bf362d-c626-4e5d-8ff4-6e847ef4eb47-catalog-content\") pod \"redhat-marketplace-cmxh5\" (UID: \"d6bf362d-c626-4e5d-8ff4-6e847ef4eb47\") " pod="openshift-marketplace/redhat-marketplace-cmxh5" Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.070104 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6bf362d-c626-4e5d-8ff4-6e847ef4eb47-utilities\") pod \"redhat-marketplace-cmxh5\" (UID: \"d6bf362d-c626-4e5d-8ff4-6e847ef4eb47\") " pod="openshift-marketplace/redhat-marketplace-cmxh5" Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.088984 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vtkk\" (UniqueName: \"kubernetes.io/projected/d6bf362d-c626-4e5d-8ff4-6e847ef4eb47-kube-api-access-8vtkk\") pod \"redhat-marketplace-cmxh5\" (UID: \"d6bf362d-c626-4e5d-8ff4-6e847ef4eb47\") " pod="openshift-marketplace/redhat-marketplace-cmxh5" Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.234304 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4s8kn"] Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.235450 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4s8kn" Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.243260 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4s8kn"] Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.332185 4896 generic.go:334] "Generic (PLEG): container finished" podID="96fbf244-90e4-4ffb-8dab-1ad893364032" containerID="00c69b287acdca635a7fc79198fe36b93c612a0c32a7de997735ff7da5687b18" exitCode=0 Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.332285 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cp8c2" event={"ID":"96fbf244-90e4-4ffb-8dab-1ad893364032","Type":"ContainerDied","Data":"00c69b287acdca635a7fc79198fe36b93c612a0c32a7de997735ff7da5687b18"} Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.334010 4896 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.335525 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"2edae3ebf91d0d31b89623769aa270059f7aa06584a13011558d5cf4c4eb04f9"} Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.336851 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"54a3c18fcfb32a28df33d4ec846aa08503715a1c6a7928b06ede6e7a08c28194"} Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.338351 4896 generic.go:334] "Generic (PLEG): container finished" podID="cdc2c09e-a873-4f65-8a5f-949b4562cb10" containerID="f2f7bd8ffa24dec4c9003fe1d734877149cf693636e2be84537e9fc3a1331be9" exitCode=0 Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.338396 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7s2vj" event={"ID":"cdc2c09e-a873-4f65-8a5f-949b4562cb10","Type":"ContainerDied","Data":"f2f7bd8ffa24dec4c9003fe1d734877149cf693636e2be84537e9fc3a1331be9"} Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.338412 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7s2vj" event={"ID":"cdc2c09e-a873-4f65-8a5f-949b4562cb10","Type":"ContainerStarted","Data":"3034d6dfd58bd6011dbc42c2834cf838b0facec47301f3048de9931fd7a695bb"} Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.340155 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" event={"ID":"351aa26e-939a-4b69-a7ee-55be7111ee6e","Type":"ContainerStarted","Data":"ab354663dabb4d91e61c157af9994bb2c779a26ba987a35acf4e43c3b4490abd"} Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.340177 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.340187 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" event={"ID":"351aa26e-939a-4b69-a7ee-55be7111ee6e","Type":"ContainerStarted","Data":"d0f8c759aecf10320a99df9dc665d78d9c1a89a4bab29001d2ab4eeaeb6b7807"} Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.341438 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"52b1f432-3929-4d34-aa46-ea75deb524be","Type":"ContainerStarted","Data":"79684100feeaf1c71af60ee0c0680c268a0d45152e8abdc94b83a8ccf3f86240"} Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.341461 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"52b1f432-3929-4d34-aa46-ea75deb524be","Type":"ContainerStarted","Data":"905383d6259d1d822dd22bfc09532dec5eba585fabb3b6dad652ebcee0bf99be"} Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.343061 4896 generic.go:334] "Generic (PLEG): container finished" podID="f48b5c5e-4ec2-4bca-bd26-02b837c029ea" containerID="87db43e7f5035089862903966e8817b18224516d05c374ea3d257cd62f64b3c9" exitCode=0 Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.343110 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wzwx5" event={"ID":"f48b5c5e-4ec2-4bca-bd26-02b837c029ea","Type":"ContainerDied","Data":"87db43e7f5035089862903966e8817b18224516d05c374ea3d257cd62f64b3c9"} Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.343128 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wzwx5" event={"ID":"f48b5c5e-4ec2-4bca-bd26-02b837c029ea","Type":"ContainerStarted","Data":"a1c7b8fcec9fe0141386698656debe691d33b9e1f90570f42b0cd7a244cd9fde"} Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.343713 4896 patch_prober.go:28] interesting pod/router-default-5444994796-8cq46 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 14:57:33 crc kubenswrapper[4896]: [-]has-synced failed: reason withheld Feb 18 14:57:33 crc kubenswrapper[4896]: [+]process-running ok Feb 18 14:57:33 crc kubenswrapper[4896]: healthz check failed Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.343746 4896 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8cq46" podUID="7a51266e-326e-4b5c-a25a-fb61bf1f3757" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.347351 4896 generic.go:334] "Generic (PLEG): container finished" podID="aa56560c-aab8-481c-9e3e-ac97ed9b69a5" containerID="9b43efaeef2cc38cc545122e5ca4054ccb0509bf72f9d9341956d94094df65aa" exitCode=0 Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.348732 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-46f87" event={"ID":"aa56560c-aab8-481c-9e3e-ac97ed9b69a5","Type":"ContainerDied","Data":"9b43efaeef2cc38cc545122e5ca4054ccb0509bf72f9d9341956d94094df65aa"} Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.374626 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cmxh5" Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.375117 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c93c74b9-3b90-4f6d-ab3b-8210d0704ce0-utilities\") pod \"redhat-marketplace-4s8kn\" (UID: \"c93c74b9-3b90-4f6d-ab3b-8210d0704ce0\") " pod="openshift-marketplace/redhat-marketplace-4s8kn" Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.375187 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rp6r5\" (UniqueName: \"kubernetes.io/projected/c93c74b9-3b90-4f6d-ab3b-8210d0704ce0-kube-api-access-rp6r5\") pod \"redhat-marketplace-4s8kn\" (UID: \"c93c74b9-3b90-4f6d-ab3b-8210d0704ce0\") " pod="openshift-marketplace/redhat-marketplace-4s8kn" Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.375281 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c93c74b9-3b90-4f6d-ab3b-8210d0704ce0-catalog-content\") pod \"redhat-marketplace-4s8kn\" (UID: \"c93c74b9-3b90-4f6d-ab3b-8210d0704ce0\") " pod="openshift-marketplace/redhat-marketplace-4s8kn" Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.449691 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.449671971 podStartE2EDuration="2.449671971s" podCreationTimestamp="2026-02-18 14:57:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:33.444590293 +0000 UTC m=+152.373924506" watchObservedRunningTime="2026-02-18 14:57:33.449671971 +0000 UTC m=+152.379006184" Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.462864 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" podStartSLOduration=129.462848053 podStartE2EDuration="2m9.462848053s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:33.462109272 +0000 UTC m=+152.391443485" watchObservedRunningTime="2026-02-18 14:57:33.462848053 +0000 UTC m=+152.392182266" Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.478741 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c93c74b9-3b90-4f6d-ab3b-8210d0704ce0-utilities\") pod \"redhat-marketplace-4s8kn\" (UID: \"c93c74b9-3b90-4f6d-ab3b-8210d0704ce0\") " pod="openshift-marketplace/redhat-marketplace-4s8kn" Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.478784 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rp6r5\" (UniqueName: \"kubernetes.io/projected/c93c74b9-3b90-4f6d-ab3b-8210d0704ce0-kube-api-access-rp6r5\") pod \"redhat-marketplace-4s8kn\" (UID: \"c93c74b9-3b90-4f6d-ab3b-8210d0704ce0\") " pod="openshift-marketplace/redhat-marketplace-4s8kn" Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.478876 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c93c74b9-3b90-4f6d-ab3b-8210d0704ce0-catalog-content\") pod \"redhat-marketplace-4s8kn\" (UID: \"c93c74b9-3b90-4f6d-ab3b-8210d0704ce0\") " pod="openshift-marketplace/redhat-marketplace-4s8kn" Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.479570 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c93c74b9-3b90-4f6d-ab3b-8210d0704ce0-utilities\") pod \"redhat-marketplace-4s8kn\" (UID: \"c93c74b9-3b90-4f6d-ab3b-8210d0704ce0\") " pod="openshift-marketplace/redhat-marketplace-4s8kn" Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.482594 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c93c74b9-3b90-4f6d-ab3b-8210d0704ce0-catalog-content\") pod \"redhat-marketplace-4s8kn\" (UID: \"c93c74b9-3b90-4f6d-ab3b-8210d0704ce0\") " pod="openshift-marketplace/redhat-marketplace-4s8kn" Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.501066 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rp6r5\" (UniqueName: \"kubernetes.io/projected/c93c74b9-3b90-4f6d-ab3b-8210d0704ce0-kube-api-access-rp6r5\") pod \"redhat-marketplace-4s8kn\" (UID: \"c93c74b9-3b90-4f6d-ab3b-8210d0704ce0\") " pod="openshift-marketplace/redhat-marketplace-4s8kn" Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.561181 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4s8kn" Feb 18 14:57:33 crc kubenswrapper[4896]: I0218 14:57:33.824591 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cmxh5"] Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.019762 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4s8kn"] Feb 18 14:57:34 crc kubenswrapper[4896]: W0218 14:57:34.029427 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc93c74b9_3b90_4f6d_ab3b_8210d0704ce0.slice/crio-b17b44a752c005757e232bc39e4792bcee43fb6a109124a55311dc9357079bac WatchSource:0}: Error finding container b17b44a752c005757e232bc39e4792bcee43fb6a109124a55311dc9357079bac: Status 404 returned error can't find the container with id b17b44a752c005757e232bc39e4792bcee43fb6a109124a55311dc9357079bac Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.076509 4896 patch_prober.go:28] interesting pod/machine-config-daemon-24rq5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.076590 4896 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.241068 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gq2n2"] Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.242516 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gq2n2" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.244983 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.245738 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gq2n2"] Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.269286 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-p7zgl" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.330371 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.331266 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.335108 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.342903 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-8cq46" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.342933 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.343247 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.344186 4896 patch_prober.go:28] interesting pod/router-default-5444994796-8cq46 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 14:57:34 crc kubenswrapper[4896]: [-]has-synced failed: reason withheld Feb 18 14:57:34 crc kubenswrapper[4896]: [+]process-running ok Feb 18 14:57:34 crc kubenswrapper[4896]: healthz check failed Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.344226 4896 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8cq46" podUID="7a51266e-326e-4b5c-a25a-fb61bf1f3757" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.372266 4896 generic.go:334] "Generic (PLEG): container finished" podID="d6bf362d-c626-4e5d-8ff4-6e847ef4eb47" containerID="722f7e38c7392b6614356182a05a0dccd0dfa1c4f827fcaf02d7d4e95ff0ca8d" exitCode=0 Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.372332 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cmxh5" event={"ID":"d6bf362d-c626-4e5d-8ff4-6e847ef4eb47","Type":"ContainerDied","Data":"722f7e38c7392b6614356182a05a0dccd0dfa1c4f827fcaf02d7d4e95ff0ca8d"} Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.372355 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cmxh5" event={"ID":"d6bf362d-c626-4e5d-8ff4-6e847ef4eb47","Type":"ContainerStarted","Data":"214acc65ab155c5b608cc8d7946d0c061bdff7ef7c802d36193835c716eae06f"} Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.376328 4896 generic.go:334] "Generic (PLEG): container finished" podID="c93c74b9-3b90-4f6d-ab3b-8210d0704ce0" containerID="847bb87210071ac7eb84aae032f534274e27b06285ddd17a55e3e381c7e7a989" exitCode=0 Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.376393 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4s8kn" event={"ID":"c93c74b9-3b90-4f6d-ab3b-8210d0704ce0","Type":"ContainerDied","Data":"847bb87210071ac7eb84aae032f534274e27b06285ddd17a55e3e381c7e7a989"} Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.376421 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4s8kn" event={"ID":"c93c74b9-3b90-4f6d-ab3b-8210d0704ce0","Type":"ContainerStarted","Data":"b17b44a752c005757e232bc39e4792bcee43fb6a109124a55311dc9357079bac"} Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.385784 4896 generic.go:334] "Generic (PLEG): container finished" podID="52b1f432-3929-4d34-aa46-ea75deb524be" containerID="79684100feeaf1c71af60ee0c0680c268a0d45152e8abdc94b83a8ccf3f86240" exitCode=0 Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.386331 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"52b1f432-3929-4d34-aa46-ea75deb524be","Type":"ContainerDied","Data":"79684100feeaf1c71af60ee0c0680c268a0d45152e8abdc94b83a8ccf3f86240"} Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.399152 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hd44s\" (UniqueName: \"kubernetes.io/projected/b8604b65-3d0c-40e3-93bb-00ff1169f8f9-kube-api-access-hd44s\") pod \"redhat-operators-gq2n2\" (UID: \"b8604b65-3d0c-40e3-93bb-00ff1169f8f9\") " pod="openshift-marketplace/redhat-operators-gq2n2" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.399195 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/37e4958e-a3b9-4165-adbe-4db4c227778d-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"37e4958e-a3b9-4165-adbe-4db4c227778d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.399309 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8604b65-3d0c-40e3-93bb-00ff1169f8f9-utilities\") pod \"redhat-operators-gq2n2\" (UID: \"b8604b65-3d0c-40e3-93bb-00ff1169f8f9\") " pod="openshift-marketplace/redhat-operators-gq2n2" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.399386 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/37e4958e-a3b9-4165-adbe-4db4c227778d-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"37e4958e-a3b9-4165-adbe-4db4c227778d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.399410 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8604b65-3d0c-40e3-93bb-00ff1169f8f9-catalog-content\") pod \"redhat-operators-gq2n2\" (UID: \"b8604b65-3d0c-40e3-93bb-00ff1169f8f9\") " pod="openshift-marketplace/redhat-operators-gq2n2" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.500310 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hd44s\" (UniqueName: \"kubernetes.io/projected/b8604b65-3d0c-40e3-93bb-00ff1169f8f9-kube-api-access-hd44s\") pod \"redhat-operators-gq2n2\" (UID: \"b8604b65-3d0c-40e3-93bb-00ff1169f8f9\") " pod="openshift-marketplace/redhat-operators-gq2n2" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.500358 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/37e4958e-a3b9-4165-adbe-4db4c227778d-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"37e4958e-a3b9-4165-adbe-4db4c227778d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.500487 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8604b65-3d0c-40e3-93bb-00ff1169f8f9-utilities\") pod \"redhat-operators-gq2n2\" (UID: \"b8604b65-3d0c-40e3-93bb-00ff1169f8f9\") " pod="openshift-marketplace/redhat-operators-gq2n2" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.500570 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/37e4958e-a3b9-4165-adbe-4db4c227778d-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"37e4958e-a3b9-4165-adbe-4db4c227778d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.500606 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8604b65-3d0c-40e3-93bb-00ff1169f8f9-catalog-content\") pod \"redhat-operators-gq2n2\" (UID: \"b8604b65-3d0c-40e3-93bb-00ff1169f8f9\") " pod="openshift-marketplace/redhat-operators-gq2n2" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.501019 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8604b65-3d0c-40e3-93bb-00ff1169f8f9-catalog-content\") pod \"redhat-operators-gq2n2\" (UID: \"b8604b65-3d0c-40e3-93bb-00ff1169f8f9\") " pod="openshift-marketplace/redhat-operators-gq2n2" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.501083 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/37e4958e-a3b9-4165-adbe-4db4c227778d-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"37e4958e-a3b9-4165-adbe-4db4c227778d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.501140 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8604b65-3d0c-40e3-93bb-00ff1169f8f9-utilities\") pod \"redhat-operators-gq2n2\" (UID: \"b8604b65-3d0c-40e3-93bb-00ff1169f8f9\") " pod="openshift-marketplace/redhat-operators-gq2n2" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.521885 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/37e4958e-a3b9-4165-adbe-4db4c227778d-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"37e4958e-a3b9-4165-adbe-4db4c227778d\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.524171 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hd44s\" (UniqueName: \"kubernetes.io/projected/b8604b65-3d0c-40e3-93bb-00ff1169f8f9-kube-api-access-hd44s\") pod \"redhat-operators-gq2n2\" (UID: \"b8604b65-3d0c-40e3-93bb-00ff1169f8f9\") " pod="openshift-marketplace/redhat-operators-gq2n2" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.570159 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.576873 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-tgwt9" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.581650 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gq2n2" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.660456 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.664697 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.682412 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-t6sfb"] Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.683646 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t6sfb" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.725812 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t6sfb"] Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.774647 4896 patch_prober.go:28] interesting pod/downloads-7954f5f757-6v4cn container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.774718 4896 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-6v4cn" podUID="036e7602-7608-4033-b18d-d859a7eddebd" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.774756 4896 patch_prober.go:28] interesting pod/downloads-7954f5f757-6v4cn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.774816 4896 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6v4cn" podUID="036e7602-7608-4033-b18d-d859a7eddebd" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.805119 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dc4l9\" (UniqueName: \"kubernetes.io/projected/eddbcdff-a387-4f95-9277-5850b6485036-kube-api-access-dc4l9\") pod \"redhat-operators-t6sfb\" (UID: \"eddbcdff-a387-4f95-9277-5850b6485036\") " pod="openshift-marketplace/redhat-operators-t6sfb" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.805203 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eddbcdff-a387-4f95-9277-5850b6485036-utilities\") pod \"redhat-operators-t6sfb\" (UID: \"eddbcdff-a387-4f95-9277-5850b6485036\") " pod="openshift-marketplace/redhat-operators-t6sfb" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.805268 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eddbcdff-a387-4f95-9277-5850b6485036-catalog-content\") pod \"redhat-operators-t6sfb\" (UID: \"eddbcdff-a387-4f95-9277-5850b6485036\") " pod="openshift-marketplace/redhat-operators-t6sfb" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.861616 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-27bbf" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.906905 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eddbcdff-a387-4f95-9277-5850b6485036-catalog-content\") pod \"redhat-operators-t6sfb\" (UID: \"eddbcdff-a387-4f95-9277-5850b6485036\") " pod="openshift-marketplace/redhat-operators-t6sfb" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.906989 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dc4l9\" (UniqueName: \"kubernetes.io/projected/eddbcdff-a387-4f95-9277-5850b6485036-kube-api-access-dc4l9\") pod \"redhat-operators-t6sfb\" (UID: \"eddbcdff-a387-4f95-9277-5850b6485036\") " pod="openshift-marketplace/redhat-operators-t6sfb" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.907073 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eddbcdff-a387-4f95-9277-5850b6485036-utilities\") pod \"redhat-operators-t6sfb\" (UID: \"eddbcdff-a387-4f95-9277-5850b6485036\") " pod="openshift-marketplace/redhat-operators-t6sfb" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.908205 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eddbcdff-a387-4f95-9277-5850b6485036-catalog-content\") pod \"redhat-operators-t6sfb\" (UID: \"eddbcdff-a387-4f95-9277-5850b6485036\") " pod="openshift-marketplace/redhat-operators-t6sfb" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.909813 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eddbcdff-a387-4f95-9277-5850b6485036-utilities\") pod \"redhat-operators-t6sfb\" (UID: \"eddbcdff-a387-4f95-9277-5850b6485036\") " pod="openshift-marketplace/redhat-operators-t6sfb" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.909879 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-4bxkk" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.910149 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-4bxkk" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.913441 4896 patch_prober.go:28] interesting pod/console-f9d7485db-4bxkk container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.17:8443/health\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.913496 4896 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-4bxkk" podUID="fa6c59b4-6823-47ef-a8a6-52c94e53ef4c" containerName="console" probeResult="failure" output="Get \"https://10.217.0.17:8443/health\": dial tcp 10.217.0.17:8443: connect: connection refused" Feb 18 14:57:34 crc kubenswrapper[4896]: I0218 14:57:34.992218 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dc4l9\" (UniqueName: \"kubernetes.io/projected/eddbcdff-a387-4f95-9277-5850b6485036-kube-api-access-dc4l9\") pod \"redhat-operators-t6sfb\" (UID: \"eddbcdff-a387-4f95-9277-5850b6485036\") " pod="openshift-marketplace/redhat-operators-t6sfb" Feb 18 14:57:35 crc kubenswrapper[4896]: I0218 14:57:35.011421 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t6sfb" Feb 18 14:57:35 crc kubenswrapper[4896]: I0218 14:57:35.355394 4896 patch_prober.go:28] interesting pod/router-default-5444994796-8cq46 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 14:57:35 crc kubenswrapper[4896]: [-]has-synced failed: reason withheld Feb 18 14:57:35 crc kubenswrapper[4896]: [+]process-running ok Feb 18 14:57:35 crc kubenswrapper[4896]: healthz check failed Feb 18 14:57:35 crc kubenswrapper[4896]: I0218 14:57:35.355451 4896 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8cq46" podUID="7a51266e-326e-4b5c-a25a-fb61bf1f3757" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 14:57:35 crc kubenswrapper[4896]: I0218 14:57:35.361475 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gq2n2"] Feb 18 14:57:35 crc kubenswrapper[4896]: I0218 14:57:35.417370 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gq2n2" event={"ID":"b8604b65-3d0c-40e3-93bb-00ff1169f8f9","Type":"ContainerStarted","Data":"8b3d42e307e6afbebb469295c9fad75c7bdcb8d59d59061f6a7f0c26ad8de774"} Feb 18 14:57:35 crc kubenswrapper[4896]: I0218 14:57:35.468429 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 18 14:57:35 crc kubenswrapper[4896]: I0218 14:57:35.584209 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t6sfb"] Feb 18 14:57:35 crc kubenswrapper[4896]: W0218 14:57:35.685364 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeddbcdff_a387_4f95_9277_5850b6485036.slice/crio-ffde705eb9656db2f46d1850ecd5a7172d23eb4b4a8313ffe1f197d209e52cf4 WatchSource:0}: Error finding container ffde705eb9656db2f46d1850ecd5a7172d23eb4b4a8313ffe1f197d209e52cf4: Status 404 returned error can't find the container with id ffde705eb9656db2f46d1850ecd5a7172d23eb4b4a8313ffe1f197d209e52cf4 Feb 18 14:57:35 crc kubenswrapper[4896]: I0218 14:57:35.706296 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 18 14:57:35 crc kubenswrapper[4896]: I0218 14:57:35.823190 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/52b1f432-3929-4d34-aa46-ea75deb524be-kubelet-dir\") pod \"52b1f432-3929-4d34-aa46-ea75deb524be\" (UID: \"52b1f432-3929-4d34-aa46-ea75deb524be\") " Feb 18 14:57:35 crc kubenswrapper[4896]: I0218 14:57:35.824378 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/52b1f432-3929-4d34-aa46-ea75deb524be-kube-api-access\") pod \"52b1f432-3929-4d34-aa46-ea75deb524be\" (UID: \"52b1f432-3929-4d34-aa46-ea75deb524be\") " Feb 18 14:57:35 crc kubenswrapper[4896]: I0218 14:57:35.823360 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/52b1f432-3929-4d34-aa46-ea75deb524be-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "52b1f432-3929-4d34-aa46-ea75deb524be" (UID: "52b1f432-3929-4d34-aa46-ea75deb524be"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 14:57:35 crc kubenswrapper[4896]: I0218 14:57:35.824943 4896 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/52b1f432-3929-4d34-aa46-ea75deb524be-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 18 14:57:35 crc kubenswrapper[4896]: I0218 14:57:35.830088 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52b1f432-3929-4d34-aa46-ea75deb524be-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "52b1f432-3929-4d34-aa46-ea75deb524be" (UID: "52b1f432-3929-4d34-aa46-ea75deb524be"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:57:35 crc kubenswrapper[4896]: I0218 14:57:35.926578 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/52b1f432-3929-4d34-aa46-ea75deb524be-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 18 14:57:36 crc kubenswrapper[4896]: I0218 14:57:36.341839 4896 patch_prober.go:28] interesting pod/router-default-5444994796-8cq46 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 14:57:36 crc kubenswrapper[4896]: [-]has-synced failed: reason withheld Feb 18 14:57:36 crc kubenswrapper[4896]: [+]process-running ok Feb 18 14:57:36 crc kubenswrapper[4896]: healthz check failed Feb 18 14:57:36 crc kubenswrapper[4896]: I0218 14:57:36.342179 4896 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8cq46" podUID="7a51266e-326e-4b5c-a25a-fb61bf1f3757" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 14:57:36 crc kubenswrapper[4896]: I0218 14:57:36.431648 4896 generic.go:334] "Generic (PLEG): container finished" podID="b8604b65-3d0c-40e3-93bb-00ff1169f8f9" containerID="2b66d9b2b9040933d5dc10a17b7f7c2b6f687fec249eb138255ce8952a4b29e5" exitCode=0 Feb 18 14:57:36 crc kubenswrapper[4896]: I0218 14:57:36.431734 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gq2n2" event={"ID":"b8604b65-3d0c-40e3-93bb-00ff1169f8f9","Type":"ContainerDied","Data":"2b66d9b2b9040933d5dc10a17b7f7c2b6f687fec249eb138255ce8952a4b29e5"} Feb 18 14:57:36 crc kubenswrapper[4896]: I0218 14:57:36.435744 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"52b1f432-3929-4d34-aa46-ea75deb524be","Type":"ContainerDied","Data":"905383d6259d1d822dd22bfc09532dec5eba585fabb3b6dad652ebcee0bf99be"} Feb 18 14:57:36 crc kubenswrapper[4896]: I0218 14:57:36.435778 4896 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="905383d6259d1d822dd22bfc09532dec5eba585fabb3b6dad652ebcee0bf99be" Feb 18 14:57:36 crc kubenswrapper[4896]: I0218 14:57:36.435782 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 18 14:57:36 crc kubenswrapper[4896]: I0218 14:57:36.451811 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"37e4958e-a3b9-4165-adbe-4db4c227778d","Type":"ContainerStarted","Data":"5e54fab1928b2df1f8011487371d10576f45de0f4db6479b1c3b48b9acd40679"} Feb 18 14:57:36 crc kubenswrapper[4896]: I0218 14:57:36.451856 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"37e4958e-a3b9-4165-adbe-4db4c227778d","Type":"ContainerStarted","Data":"5eb6c7a46c578933714da4659c91e31fafc9ba89e04769ca269e5eda070bfc02"} Feb 18 14:57:36 crc kubenswrapper[4896]: I0218 14:57:36.458467 4896 generic.go:334] "Generic (PLEG): container finished" podID="eddbcdff-a387-4f95-9277-5850b6485036" containerID="9bc72d58d15b9d44e16885e79087167612fa6753010eba7d950dd37efe2cf549" exitCode=0 Feb 18 14:57:36 crc kubenswrapper[4896]: I0218 14:57:36.458529 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t6sfb" event={"ID":"eddbcdff-a387-4f95-9277-5850b6485036","Type":"ContainerDied","Data":"9bc72d58d15b9d44e16885e79087167612fa6753010eba7d950dd37efe2cf549"} Feb 18 14:57:36 crc kubenswrapper[4896]: I0218 14:57:36.458552 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t6sfb" event={"ID":"eddbcdff-a387-4f95-9277-5850b6485036","Type":"ContainerStarted","Data":"ffde705eb9656db2f46d1850ecd5a7172d23eb4b4a8313ffe1f197d209e52cf4"} Feb 18 14:57:36 crc kubenswrapper[4896]: I0218 14:57:36.470084 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.470065341 podStartE2EDuration="2.470065341s" podCreationTimestamp="2026-02-18 14:57:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:57:36.469111793 +0000 UTC m=+155.398446006" watchObservedRunningTime="2026-02-18 14:57:36.470065341 +0000 UTC m=+155.399399554" Feb 18 14:57:36 crc kubenswrapper[4896]: I0218 14:57:36.646319 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 14:57:36 crc kubenswrapper[4896]: I0218 14:57:36.784930 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-5795c" Feb 18 14:57:37 crc kubenswrapper[4896]: I0218 14:57:37.341755 4896 patch_prober.go:28] interesting pod/router-default-5444994796-8cq46 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 14:57:37 crc kubenswrapper[4896]: [-]has-synced failed: reason withheld Feb 18 14:57:37 crc kubenswrapper[4896]: [+]process-running ok Feb 18 14:57:37 crc kubenswrapper[4896]: healthz check failed Feb 18 14:57:37 crc kubenswrapper[4896]: I0218 14:57:37.341832 4896 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8cq46" podUID="7a51266e-326e-4b5c-a25a-fb61bf1f3757" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 14:57:37 crc kubenswrapper[4896]: I0218 14:57:37.468968 4896 generic.go:334] "Generic (PLEG): container finished" podID="37e4958e-a3b9-4165-adbe-4db4c227778d" containerID="5e54fab1928b2df1f8011487371d10576f45de0f4db6479b1c3b48b9acd40679" exitCode=0 Feb 18 14:57:37 crc kubenswrapper[4896]: I0218 14:57:37.469018 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"37e4958e-a3b9-4165-adbe-4db4c227778d","Type":"ContainerDied","Data":"5e54fab1928b2df1f8011487371d10576f45de0f4db6479b1c3b48b9acd40679"} Feb 18 14:57:38 crc kubenswrapper[4896]: I0218 14:57:38.339195 4896 patch_prober.go:28] interesting pod/router-default-5444994796-8cq46 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 14:57:38 crc kubenswrapper[4896]: [-]has-synced failed: reason withheld Feb 18 14:57:38 crc kubenswrapper[4896]: [+]process-running ok Feb 18 14:57:38 crc kubenswrapper[4896]: healthz check failed Feb 18 14:57:38 crc kubenswrapper[4896]: I0218 14:57:38.339610 4896 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8cq46" podUID="7a51266e-326e-4b5c-a25a-fb61bf1f3757" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 14:57:38 crc kubenswrapper[4896]: I0218 14:57:38.984643 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 18 14:57:39 crc kubenswrapper[4896]: I0218 14:57:39.098261 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/37e4958e-a3b9-4165-adbe-4db4c227778d-kubelet-dir\") pod \"37e4958e-a3b9-4165-adbe-4db4c227778d\" (UID: \"37e4958e-a3b9-4165-adbe-4db4c227778d\") " Feb 18 14:57:39 crc kubenswrapper[4896]: I0218 14:57:39.098330 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/37e4958e-a3b9-4165-adbe-4db4c227778d-kube-api-access\") pod \"37e4958e-a3b9-4165-adbe-4db4c227778d\" (UID: \"37e4958e-a3b9-4165-adbe-4db4c227778d\") " Feb 18 14:57:39 crc kubenswrapper[4896]: I0218 14:57:39.098403 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/37e4958e-a3b9-4165-adbe-4db4c227778d-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "37e4958e-a3b9-4165-adbe-4db4c227778d" (UID: "37e4958e-a3b9-4165-adbe-4db4c227778d"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 14:57:39 crc kubenswrapper[4896]: I0218 14:57:39.098623 4896 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/37e4958e-a3b9-4165-adbe-4db4c227778d-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 18 14:57:39 crc kubenswrapper[4896]: I0218 14:57:39.106062 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37e4958e-a3b9-4165-adbe-4db4c227778d-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "37e4958e-a3b9-4165-adbe-4db4c227778d" (UID: "37e4958e-a3b9-4165-adbe-4db4c227778d"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:57:39 crc kubenswrapper[4896]: I0218 14:57:39.200121 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/37e4958e-a3b9-4165-adbe-4db4c227778d-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 18 14:57:39 crc kubenswrapper[4896]: I0218 14:57:39.340389 4896 patch_prober.go:28] interesting pod/router-default-5444994796-8cq46 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 14:57:39 crc kubenswrapper[4896]: [-]has-synced failed: reason withheld Feb 18 14:57:39 crc kubenswrapper[4896]: [+]process-running ok Feb 18 14:57:39 crc kubenswrapper[4896]: healthz check failed Feb 18 14:57:39 crc kubenswrapper[4896]: I0218 14:57:39.340449 4896 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8cq46" podUID="7a51266e-326e-4b5c-a25a-fb61bf1f3757" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 14:57:39 crc kubenswrapper[4896]: I0218 14:57:39.496360 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"37e4958e-a3b9-4165-adbe-4db4c227778d","Type":"ContainerDied","Data":"5eb6c7a46c578933714da4659c91e31fafc9ba89e04769ca269e5eda070bfc02"} Feb 18 14:57:39 crc kubenswrapper[4896]: I0218 14:57:39.496408 4896 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5eb6c7a46c578933714da4659c91e31fafc9ba89e04769ca269e5eda070bfc02" Feb 18 14:57:39 crc kubenswrapper[4896]: I0218 14:57:39.496544 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 18 14:57:40 crc kubenswrapper[4896]: I0218 14:57:40.340436 4896 patch_prober.go:28] interesting pod/router-default-5444994796-8cq46 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 14:57:40 crc kubenswrapper[4896]: [-]has-synced failed: reason withheld Feb 18 14:57:40 crc kubenswrapper[4896]: [+]process-running ok Feb 18 14:57:40 crc kubenswrapper[4896]: healthz check failed Feb 18 14:57:40 crc kubenswrapper[4896]: I0218 14:57:40.340511 4896 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8cq46" podUID="7a51266e-326e-4b5c-a25a-fb61bf1f3757" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 14:57:41 crc kubenswrapper[4896]: I0218 14:57:41.338867 4896 patch_prober.go:28] interesting pod/router-default-5444994796-8cq46 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 14:57:41 crc kubenswrapper[4896]: [-]has-synced failed: reason withheld Feb 18 14:57:41 crc kubenswrapper[4896]: [+]process-running ok Feb 18 14:57:41 crc kubenswrapper[4896]: healthz check failed Feb 18 14:57:41 crc kubenswrapper[4896]: I0218 14:57:41.339204 4896 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8cq46" podUID="7a51266e-326e-4b5c-a25a-fb61bf1f3757" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 14:57:42 crc kubenswrapper[4896]: I0218 14:57:42.339613 4896 patch_prober.go:28] interesting pod/router-default-5444994796-8cq46 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 18 14:57:42 crc kubenswrapper[4896]: [-]has-synced failed: reason withheld Feb 18 14:57:42 crc kubenswrapper[4896]: [+]process-running ok Feb 18 14:57:42 crc kubenswrapper[4896]: healthz check failed Feb 18 14:57:42 crc kubenswrapper[4896]: I0218 14:57:42.339680 4896 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-8cq46" podUID="7a51266e-326e-4b5c-a25a-fb61bf1f3757" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 18 14:57:43 crc kubenswrapper[4896]: I0218 14:57:43.339562 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-8cq46" Feb 18 14:57:43 crc kubenswrapper[4896]: I0218 14:57:43.343368 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-8cq46" Feb 18 14:57:44 crc kubenswrapper[4896]: I0218 14:57:44.756544 4896 patch_prober.go:28] interesting pod/downloads-7954f5f757-6v4cn container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Feb 18 14:57:44 crc kubenswrapper[4896]: I0218 14:57:44.756627 4896 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-6v4cn" podUID="036e7602-7608-4033-b18d-d859a7eddebd" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Feb 18 14:57:44 crc kubenswrapper[4896]: I0218 14:57:44.757335 4896 patch_prober.go:28] interesting pod/downloads-7954f5f757-6v4cn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Feb 18 14:57:44 crc kubenswrapper[4896]: I0218 14:57:44.757362 4896 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-6v4cn" podUID="036e7602-7608-4033-b18d-d859a7eddebd" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Feb 18 14:57:44 crc kubenswrapper[4896]: I0218 14:57:44.953704 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-4bxkk" Feb 18 14:57:44 crc kubenswrapper[4896]: I0218 14:57:44.958156 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-4bxkk" Feb 18 14:57:46 crc kubenswrapper[4896]: I0218 14:57:46.407805 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/229213f7-7a25-4982-a449-0122912ba05c-metrics-certs\") pod \"network-metrics-daemon-98x74\" (UID: \"229213f7-7a25-4982-a449-0122912ba05c\") " pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:57:46 crc kubenswrapper[4896]: I0218 14:57:46.413658 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/229213f7-7a25-4982-a449-0122912ba05c-metrics-certs\") pod \"network-metrics-daemon-98x74\" (UID: \"229213f7-7a25-4982-a449-0122912ba05c\") " pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:57:46 crc kubenswrapper[4896]: I0218 14:57:46.652979 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-98x74" Feb 18 14:57:50 crc kubenswrapper[4896]: I0218 14:57:50.549169 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-27bbf"] Feb 18 14:57:50 crc kubenswrapper[4896]: I0218 14:57:50.549445 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-27bbf" podUID="31b0687b-5c42-4259-b96d-284c1aa487b4" containerName="controller-manager" containerID="cri-o://6709c0c6bc6eb8fad3c1e2c00d5547edbd703aa0fbbbea4809220b10b020f485" gracePeriod=30 Feb 18 14:57:50 crc kubenswrapper[4896]: I0218 14:57:50.565702 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-v6thv"] Feb 18 14:57:50 crc kubenswrapper[4896]: I0218 14:57:50.565893 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v6thv" podUID="0c6cb779-d660-45f8-8fb2-5829d42ee030" containerName="route-controller-manager" containerID="cri-o://48efb299ab8ed41e49575c9749f3da5efa5277019eabe5db6fe02dcf8d920bd0" gracePeriod=30 Feb 18 14:57:51 crc kubenswrapper[4896]: I0218 14:57:51.556004 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 14:57:51 crc kubenswrapper[4896]: I0218 14:57:51.593284 4896 generic.go:334] "Generic (PLEG): container finished" podID="31b0687b-5c42-4259-b96d-284c1aa487b4" containerID="6709c0c6bc6eb8fad3c1e2c00d5547edbd703aa0fbbbea4809220b10b020f485" exitCode=0 Feb 18 14:57:51 crc kubenswrapper[4896]: I0218 14:57:51.593370 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-27bbf" event={"ID":"31b0687b-5c42-4259-b96d-284c1aa487b4","Type":"ContainerDied","Data":"6709c0c6bc6eb8fad3c1e2c00d5547edbd703aa0fbbbea4809220b10b020f485"} Feb 18 14:57:51 crc kubenswrapper[4896]: I0218 14:57:51.597008 4896 generic.go:334] "Generic (PLEG): container finished" podID="0c6cb779-d660-45f8-8fb2-5829d42ee030" containerID="48efb299ab8ed41e49575c9749f3da5efa5277019eabe5db6fe02dcf8d920bd0" exitCode=0 Feb 18 14:57:51 crc kubenswrapper[4896]: I0218 14:57:51.597214 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v6thv" event={"ID":"0c6cb779-d660-45f8-8fb2-5829d42ee030","Type":"ContainerDied","Data":"48efb299ab8ed41e49575c9749f3da5efa5277019eabe5db6fe02dcf8d920bd0"} Feb 18 14:57:53 crc kubenswrapper[4896]: I0218 14:57:53.473671 4896 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-v6thv container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Feb 18 14:57:53 crc kubenswrapper[4896]: I0218 14:57:53.473726 4896 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v6thv" podUID="0c6cb779-d660-45f8-8fb2-5829d42ee030" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Feb 18 14:57:54 crc kubenswrapper[4896]: I0218 14:57:54.761334 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-6v4cn" Feb 18 14:57:54 crc kubenswrapper[4896]: I0218 14:57:54.845362 4896 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-27bbf container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Feb 18 14:57:54 crc kubenswrapper[4896]: I0218 14:57:54.845419 4896 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-27bbf" podUID="31b0687b-5c42-4259-b96d-284c1aa487b4" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Feb 18 14:58:03 crc kubenswrapper[4896]: I0218 14:58:03.474146 4896 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-v6thv container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Feb 18 14:58:03 crc kubenswrapper[4896]: I0218 14:58:03.474796 4896 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v6thv" podUID="0c6cb779-d660-45f8-8fb2-5829d42ee030" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Feb 18 14:58:04 crc kubenswrapper[4896]: I0218 14:58:04.075081 4896 patch_prober.go:28] interesting pod/machine-config-daemon-24rq5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 14:58:04 crc kubenswrapper[4896]: I0218 14:58:04.075396 4896 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 14:58:04 crc kubenswrapper[4896]: I0218 14:58:04.845958 4896 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-27bbf container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Feb 18 14:58:04 crc kubenswrapper[4896]: I0218 14:58:04.846046 4896 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-27bbf" podUID="31b0687b-5c42-4259-b96d-284c1aa487b4" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Feb 18 14:58:05 crc kubenswrapper[4896]: I0218 14:58:05.007909 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hv6vt" Feb 18 14:58:10 crc kubenswrapper[4896]: I0218 14:58:10.521910 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 18 14:58:10 crc kubenswrapper[4896]: E0218 14:58:10.524068 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52b1f432-3929-4d34-aa46-ea75deb524be" containerName="pruner" Feb 18 14:58:10 crc kubenswrapper[4896]: I0218 14:58:10.524282 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="52b1f432-3929-4d34-aa46-ea75deb524be" containerName="pruner" Feb 18 14:58:10 crc kubenswrapper[4896]: E0218 14:58:10.524460 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37e4958e-a3b9-4165-adbe-4db4c227778d" containerName="pruner" Feb 18 14:58:10 crc kubenswrapper[4896]: I0218 14:58:10.524585 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="37e4958e-a3b9-4165-adbe-4db4c227778d" containerName="pruner" Feb 18 14:58:10 crc kubenswrapper[4896]: I0218 14:58:10.524877 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="37e4958e-a3b9-4165-adbe-4db4c227778d" containerName="pruner" Feb 18 14:58:10 crc kubenswrapper[4896]: I0218 14:58:10.525172 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="52b1f432-3929-4d34-aa46-ea75deb524be" containerName="pruner" Feb 18 14:58:10 crc kubenswrapper[4896]: I0218 14:58:10.525934 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 18 14:58:10 crc kubenswrapper[4896]: I0218 14:58:10.526313 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 18 14:58:10 crc kubenswrapper[4896]: I0218 14:58:10.530908 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 18 14:58:10 crc kubenswrapper[4896]: I0218 14:58:10.531858 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 18 14:58:10 crc kubenswrapper[4896]: I0218 14:58:10.624536 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dfcee59f-69ac-42ac-86e8-20c08066d0c1-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"dfcee59f-69ac-42ac-86e8-20c08066d0c1\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 18 14:58:10 crc kubenswrapper[4896]: I0218 14:58:10.624658 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dfcee59f-69ac-42ac-86e8-20c08066d0c1-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"dfcee59f-69ac-42ac-86e8-20c08066d0c1\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 18 14:58:10 crc kubenswrapper[4896]: I0218 14:58:10.726361 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dfcee59f-69ac-42ac-86e8-20c08066d0c1-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"dfcee59f-69ac-42ac-86e8-20c08066d0c1\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 18 14:58:10 crc kubenswrapper[4896]: I0218 14:58:10.726760 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dfcee59f-69ac-42ac-86e8-20c08066d0c1-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"dfcee59f-69ac-42ac-86e8-20c08066d0c1\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 18 14:58:10 crc kubenswrapper[4896]: I0218 14:58:10.726829 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dfcee59f-69ac-42ac-86e8-20c08066d0c1-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"dfcee59f-69ac-42ac-86e8-20c08066d0c1\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 18 14:58:10 crc kubenswrapper[4896]: I0218 14:58:10.959753 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 18 14:58:10 crc kubenswrapper[4896]: I0218 14:58:10.961477 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dfcee59f-69ac-42ac-86e8-20c08066d0c1-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"dfcee59f-69ac-42ac-86e8-20c08066d0c1\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 18 14:58:11 crc kubenswrapper[4896]: I0218 14:58:11.042863 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 18 14:58:14 crc kubenswrapper[4896]: I0218 14:58:14.474042 4896 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-v6thv container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: i/o timeout" start-of-body= Feb 18 14:58:14 crc kubenswrapper[4896]: I0218 14:58:14.474419 4896 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v6thv" podUID="0c6cb779-d660-45f8-8fb2-5829d42ee030" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: i/o timeout" Feb 18 14:58:14 crc kubenswrapper[4896]: E0218 14:58:14.736944 4896 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 18 14:58:14 crc kubenswrapper[4896]: E0218 14:58:14.737140 4896 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hd44s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-gq2n2_openshift-marketplace(b8604b65-3d0c-40e3-93bb-00ff1169f8f9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 18 14:58:14 crc kubenswrapper[4896]: E0218 14:58:14.738629 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-gq2n2" podUID="b8604b65-3d0c-40e3-93bb-00ff1169f8f9" Feb 18 14:58:14 crc kubenswrapper[4896]: I0218 14:58:14.930940 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 18 14:58:14 crc kubenswrapper[4896]: I0218 14:58:14.931971 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 18 14:58:14 crc kubenswrapper[4896]: I0218 14:58:14.937200 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 18 14:58:14 crc kubenswrapper[4896]: I0218 14:58:14.974890 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4a847ad6-23c1-43e2-b916-c97df45c2aff-kubelet-dir\") pod \"installer-9-crc\" (UID: \"4a847ad6-23c1-43e2-b916-c97df45c2aff\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 18 14:58:14 crc kubenswrapper[4896]: I0218 14:58:14.974985 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/4a847ad6-23c1-43e2-b916-c97df45c2aff-var-lock\") pod \"installer-9-crc\" (UID: \"4a847ad6-23c1-43e2-b916-c97df45c2aff\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 18 14:58:14 crc kubenswrapper[4896]: I0218 14:58:14.975012 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4a847ad6-23c1-43e2-b916-c97df45c2aff-kube-api-access\") pod \"installer-9-crc\" (UID: \"4a847ad6-23c1-43e2-b916-c97df45c2aff\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 18 14:58:15 crc kubenswrapper[4896]: E0218 14:58:15.002522 4896 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 18 14:58:15 crc kubenswrapper[4896]: E0218 14:58:15.002661 4896 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dc4l9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-t6sfb_openshift-marketplace(eddbcdff-a387-4f95-9277-5850b6485036): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 18 14:58:15 crc kubenswrapper[4896]: E0218 14:58:15.003792 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-t6sfb" podUID="eddbcdff-a387-4f95-9277-5850b6485036" Feb 18 14:58:15 crc kubenswrapper[4896]: I0218 14:58:15.076044 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/4a847ad6-23c1-43e2-b916-c97df45c2aff-var-lock\") pod \"installer-9-crc\" (UID: \"4a847ad6-23c1-43e2-b916-c97df45c2aff\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 18 14:58:15 crc kubenswrapper[4896]: I0218 14:58:15.076111 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4a847ad6-23c1-43e2-b916-c97df45c2aff-kube-api-access\") pod \"installer-9-crc\" (UID: \"4a847ad6-23c1-43e2-b916-c97df45c2aff\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 18 14:58:15 crc kubenswrapper[4896]: I0218 14:58:15.076178 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/4a847ad6-23c1-43e2-b916-c97df45c2aff-var-lock\") pod \"installer-9-crc\" (UID: \"4a847ad6-23c1-43e2-b916-c97df45c2aff\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 18 14:58:15 crc kubenswrapper[4896]: I0218 14:58:15.076710 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4a847ad6-23c1-43e2-b916-c97df45c2aff-kubelet-dir\") pod \"installer-9-crc\" (UID: \"4a847ad6-23c1-43e2-b916-c97df45c2aff\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 18 14:58:15 crc kubenswrapper[4896]: I0218 14:58:15.076790 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4a847ad6-23c1-43e2-b916-c97df45c2aff-kubelet-dir\") pod \"installer-9-crc\" (UID: \"4a847ad6-23c1-43e2-b916-c97df45c2aff\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 18 14:58:15 crc kubenswrapper[4896]: I0218 14:58:15.102339 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4a847ad6-23c1-43e2-b916-c97df45c2aff-kube-api-access\") pod \"installer-9-crc\" (UID: \"4a847ad6-23c1-43e2-b916-c97df45c2aff\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 18 14:58:15 crc kubenswrapper[4896]: I0218 14:58:15.277356 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 18 14:58:15 crc kubenswrapper[4896]: I0218 14:58:15.846314 4896 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-27bbf container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 18 14:58:15 crc kubenswrapper[4896]: I0218 14:58:15.846423 4896 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-27bbf" podUID="31b0687b-5c42-4259-b96d-284c1aa487b4" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 18 14:58:16 crc kubenswrapper[4896]: E0218 14:58:16.141414 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-gq2n2" podUID="b8604b65-3d0c-40e3-93bb-00ff1169f8f9" Feb 18 14:58:16 crc kubenswrapper[4896]: E0218 14:58:16.141749 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-t6sfb" podUID="eddbcdff-a387-4f95-9277-5850b6485036" Feb 18 14:58:16 crc kubenswrapper[4896]: E0218 14:58:16.308520 4896 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 18 14:58:16 crc kubenswrapper[4896]: E0218 14:58:16.308682 4896 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rp6r5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-4s8kn_openshift-marketplace(c93c74b9-3b90-4f6d-ab3b-8210d0704ce0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 18 14:58:16 crc kubenswrapper[4896]: E0218 14:58:16.310143 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-4s8kn" podUID="c93c74b9-3b90-4f6d-ab3b-8210d0704ce0" Feb 18 14:58:17 crc kubenswrapper[4896]: E0218 14:58:17.912145 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-4s8kn" podUID="c93c74b9-3b90-4f6d-ab3b-8210d0704ce0" Feb 18 14:58:18 crc kubenswrapper[4896]: E0218 14:58:18.234764 4896 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 18 14:58:18 crc kubenswrapper[4896]: E0218 14:58:18.234911 4896 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jrtfz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-wzwx5_openshift-marketplace(f48b5c5e-4ec2-4bca-bd26-02b837c029ea): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 18 14:58:18 crc kubenswrapper[4896]: E0218 14:58:18.236122 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-wzwx5" podUID="f48b5c5e-4ec2-4bca-bd26-02b837c029ea" Feb 18 14:58:19 crc kubenswrapper[4896]: E0218 14:58:19.907255 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-wzwx5" podUID="f48b5c5e-4ec2-4bca-bd26-02b837c029ea" Feb 18 14:58:19 crc kubenswrapper[4896]: I0218 14:58:19.989818 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v6thv" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.000118 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-27bbf" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.033880 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c887bd576-rw68b"] Feb 18 14:58:20 crc kubenswrapper[4896]: E0218 14:58:20.034129 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31b0687b-5c42-4259-b96d-284c1aa487b4" containerName="controller-manager" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.034142 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="31b0687b-5c42-4259-b96d-284c1aa487b4" containerName="controller-manager" Feb 18 14:58:20 crc kubenswrapper[4896]: E0218 14:58:20.034153 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c6cb779-d660-45f8-8fb2-5829d42ee030" containerName="route-controller-manager" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.035194 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c6cb779-d660-45f8-8fb2-5829d42ee030" containerName="route-controller-manager" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.035986 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="31b0687b-5c42-4259-b96d-284c1aa487b4" containerName="controller-manager" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.036006 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c6cb779-d660-45f8-8fb2-5829d42ee030" containerName="route-controller-manager" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.036516 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c887bd576-rw68b" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.048188 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31b0687b-5c42-4259-b96d-284c1aa487b4-config\") pod \"31b0687b-5c42-4259-b96d-284c1aa487b4\" (UID: \"31b0687b-5c42-4259-b96d-284c1aa487b4\") " Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.048278 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/31b0687b-5c42-4259-b96d-284c1aa487b4-proxy-ca-bundles\") pod \"31b0687b-5c42-4259-b96d-284c1aa487b4\" (UID: \"31b0687b-5c42-4259-b96d-284c1aa487b4\") " Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.048322 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-272j4\" (UniqueName: \"kubernetes.io/projected/31b0687b-5c42-4259-b96d-284c1aa487b4-kube-api-access-272j4\") pod \"31b0687b-5c42-4259-b96d-284c1aa487b4\" (UID: \"31b0687b-5c42-4259-b96d-284c1aa487b4\") " Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.048408 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhkrn\" (UniqueName: \"kubernetes.io/projected/0c6cb779-d660-45f8-8fb2-5829d42ee030-kube-api-access-hhkrn\") pod \"0c6cb779-d660-45f8-8fb2-5829d42ee030\" (UID: \"0c6cb779-d660-45f8-8fb2-5829d42ee030\") " Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.048440 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0c6cb779-d660-45f8-8fb2-5829d42ee030-client-ca\") pod \"0c6cb779-d660-45f8-8fb2-5829d42ee030\" (UID: \"0c6cb779-d660-45f8-8fb2-5829d42ee030\") " Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.048471 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/31b0687b-5c42-4259-b96d-284c1aa487b4-client-ca\") pod \"31b0687b-5c42-4259-b96d-284c1aa487b4\" (UID: \"31b0687b-5c42-4259-b96d-284c1aa487b4\") " Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.048514 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0c6cb779-d660-45f8-8fb2-5829d42ee030-serving-cert\") pod \"0c6cb779-d660-45f8-8fb2-5829d42ee030\" (UID: \"0c6cb779-d660-45f8-8fb2-5829d42ee030\") " Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.048557 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31b0687b-5c42-4259-b96d-284c1aa487b4-serving-cert\") pod \"31b0687b-5c42-4259-b96d-284c1aa487b4\" (UID: \"31b0687b-5c42-4259-b96d-284c1aa487b4\") " Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.048585 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c6cb779-d660-45f8-8fb2-5829d42ee030-config\") pod \"0c6cb779-d660-45f8-8fb2-5829d42ee030\" (UID: \"0c6cb779-d660-45f8-8fb2-5829d42ee030\") " Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.049550 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31b0687b-5c42-4259-b96d-284c1aa487b4-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "31b0687b-5c42-4259-b96d-284c1aa487b4" (UID: "31b0687b-5c42-4259-b96d-284c1aa487b4"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.050342 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31b0687b-5c42-4259-b96d-284c1aa487b4-client-ca" (OuterVolumeSpecName: "client-ca") pod "31b0687b-5c42-4259-b96d-284c1aa487b4" (UID: "31b0687b-5c42-4259-b96d-284c1aa487b4"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.050536 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c6cb779-d660-45f8-8fb2-5829d42ee030-config" (OuterVolumeSpecName: "config") pod "0c6cb779-d660-45f8-8fb2-5829d42ee030" (UID: "0c6cb779-d660-45f8-8fb2-5829d42ee030"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.053413 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31b0687b-5c42-4259-b96d-284c1aa487b4-config" (OuterVolumeSpecName: "config") pod "31b0687b-5c42-4259-b96d-284c1aa487b4" (UID: "31b0687b-5c42-4259-b96d-284c1aa487b4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.054110 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c6cb779-d660-45f8-8fb2-5829d42ee030-client-ca" (OuterVolumeSpecName: "client-ca") pod "0c6cb779-d660-45f8-8fb2-5829d42ee030" (UID: "0c6cb779-d660-45f8-8fb2-5829d42ee030"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.054834 4896 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31b0687b-5c42-4259-b96d-284c1aa487b4-config\") on node \"crc\" DevicePath \"\"" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.054872 4896 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/31b0687b-5c42-4259-b96d-284c1aa487b4-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.054883 4896 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0c6cb779-d660-45f8-8fb2-5829d42ee030-client-ca\") on node \"crc\" DevicePath \"\"" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.054892 4896 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/31b0687b-5c42-4259-b96d-284c1aa487b4-client-ca\") on node \"crc\" DevicePath \"\"" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.054900 4896 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c6cb779-d660-45f8-8fb2-5829d42ee030-config\") on node \"crc\" DevicePath \"\"" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.058463 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c887bd576-rw68b"] Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.059570 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c6cb779-d660-45f8-8fb2-5829d42ee030-kube-api-access-hhkrn" (OuterVolumeSpecName: "kube-api-access-hhkrn") pod "0c6cb779-d660-45f8-8fb2-5829d42ee030" (UID: "0c6cb779-d660-45f8-8fb2-5829d42ee030"). InnerVolumeSpecName "kube-api-access-hhkrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.059982 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c6cb779-d660-45f8-8fb2-5829d42ee030-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0c6cb779-d660-45f8-8fb2-5829d42ee030" (UID: "0c6cb779-d660-45f8-8fb2-5829d42ee030"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.065221 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31b0687b-5c42-4259-b96d-284c1aa487b4-kube-api-access-272j4" (OuterVolumeSpecName: "kube-api-access-272j4") pod "31b0687b-5c42-4259-b96d-284c1aa487b4" (UID: "31b0687b-5c42-4259-b96d-284c1aa487b4"). InnerVolumeSpecName "kube-api-access-272j4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.066584 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31b0687b-5c42-4259-b96d-284c1aa487b4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "31b0687b-5c42-4259-b96d-284c1aa487b4" (UID: "31b0687b-5c42-4259-b96d-284c1aa487b4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.156114 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/85c1cb1a-6343-490c-98a3-f11a52290a73-client-ca\") pod \"route-controller-manager-c887bd576-rw68b\" (UID: \"85c1cb1a-6343-490c-98a3-f11a52290a73\") " pod="openshift-route-controller-manager/route-controller-manager-c887bd576-rw68b" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.156476 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85c1cb1a-6343-490c-98a3-f11a52290a73-config\") pod \"route-controller-manager-c887bd576-rw68b\" (UID: \"85c1cb1a-6343-490c-98a3-f11a52290a73\") " pod="openshift-route-controller-manager/route-controller-manager-c887bd576-rw68b" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.156523 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49t2v\" (UniqueName: \"kubernetes.io/projected/85c1cb1a-6343-490c-98a3-f11a52290a73-kube-api-access-49t2v\") pod \"route-controller-manager-c887bd576-rw68b\" (UID: \"85c1cb1a-6343-490c-98a3-f11a52290a73\") " pod="openshift-route-controller-manager/route-controller-manager-c887bd576-rw68b" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.156550 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85c1cb1a-6343-490c-98a3-f11a52290a73-serving-cert\") pod \"route-controller-manager-c887bd576-rw68b\" (UID: \"85c1cb1a-6343-490c-98a3-f11a52290a73\") " pod="openshift-route-controller-manager/route-controller-manager-c887bd576-rw68b" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.156680 4896 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0c6cb779-d660-45f8-8fb2-5829d42ee030-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.156699 4896 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/31b0687b-5c42-4259-b96d-284c1aa487b4-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.156740 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-272j4\" (UniqueName: \"kubernetes.io/projected/31b0687b-5c42-4259-b96d-284c1aa487b4-kube-api-access-272j4\") on node \"crc\" DevicePath \"\"" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.156751 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhkrn\" (UniqueName: \"kubernetes.io/projected/0c6cb779-d660-45f8-8fb2-5829d42ee030-kube-api-access-hhkrn\") on node \"crc\" DevicePath \"\"" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.162687 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-98x74"] Feb 18 14:58:20 crc kubenswrapper[4896]: W0218 14:58:20.169577 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod229213f7_7a25_4982_a449_0122912ba05c.slice/crio-1a5369f27bfadc6a5272fb322d7d756de7870dd7fec528b00794c781fe1f94c1 WatchSource:0}: Error finding container 1a5369f27bfadc6a5272fb322d7d756de7870dd7fec528b00794c781fe1f94c1: Status 404 returned error can't find the container with id 1a5369f27bfadc6a5272fb322d7d756de7870dd7fec528b00794c781fe1f94c1 Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.258343 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/85c1cb1a-6343-490c-98a3-f11a52290a73-client-ca\") pod \"route-controller-manager-c887bd576-rw68b\" (UID: \"85c1cb1a-6343-490c-98a3-f11a52290a73\") " pod="openshift-route-controller-manager/route-controller-manager-c887bd576-rw68b" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.258397 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85c1cb1a-6343-490c-98a3-f11a52290a73-config\") pod \"route-controller-manager-c887bd576-rw68b\" (UID: \"85c1cb1a-6343-490c-98a3-f11a52290a73\") " pod="openshift-route-controller-manager/route-controller-manager-c887bd576-rw68b" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.258417 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49t2v\" (UniqueName: \"kubernetes.io/projected/85c1cb1a-6343-490c-98a3-f11a52290a73-kube-api-access-49t2v\") pod \"route-controller-manager-c887bd576-rw68b\" (UID: \"85c1cb1a-6343-490c-98a3-f11a52290a73\") " pod="openshift-route-controller-manager/route-controller-manager-c887bd576-rw68b" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.258440 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85c1cb1a-6343-490c-98a3-f11a52290a73-serving-cert\") pod \"route-controller-manager-c887bd576-rw68b\" (UID: \"85c1cb1a-6343-490c-98a3-f11a52290a73\") " pod="openshift-route-controller-manager/route-controller-manager-c887bd576-rw68b" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.260084 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/85c1cb1a-6343-490c-98a3-f11a52290a73-client-ca\") pod \"route-controller-manager-c887bd576-rw68b\" (UID: \"85c1cb1a-6343-490c-98a3-f11a52290a73\") " pod="openshift-route-controller-manager/route-controller-manager-c887bd576-rw68b" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.261017 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85c1cb1a-6343-490c-98a3-f11a52290a73-config\") pod \"route-controller-manager-c887bd576-rw68b\" (UID: \"85c1cb1a-6343-490c-98a3-f11a52290a73\") " pod="openshift-route-controller-manager/route-controller-manager-c887bd576-rw68b" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.265026 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85c1cb1a-6343-490c-98a3-f11a52290a73-serving-cert\") pod \"route-controller-manager-c887bd576-rw68b\" (UID: \"85c1cb1a-6343-490c-98a3-f11a52290a73\") " pod="openshift-route-controller-manager/route-controller-manager-c887bd576-rw68b" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.277011 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49t2v\" (UniqueName: \"kubernetes.io/projected/85c1cb1a-6343-490c-98a3-f11a52290a73-kube-api-access-49t2v\") pod \"route-controller-manager-c887bd576-rw68b\" (UID: \"85c1cb1a-6343-490c-98a3-f11a52290a73\") " pod="openshift-route-controller-manager/route-controller-manager-c887bd576-rw68b" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.354862 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c887bd576-rw68b" Feb 18 14:58:20 crc kubenswrapper[4896]: E0218 14:58:20.373329 4896 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 18 14:58:20 crc kubenswrapper[4896]: E0218 14:58:20.373478 4896 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gv76m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-46f87_openshift-marketplace(aa56560c-aab8-481c-9e3e-ac97ed9b69a5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 18 14:58:20 crc kubenswrapper[4896]: E0218 14:58:20.374659 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-46f87" podUID="aa56560c-aab8-481c-9e3e-ac97ed9b69a5" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.471292 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.472548 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.557498 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c887bd576-rw68b"] Feb 18 14:58:20 crc kubenswrapper[4896]: W0218 14:58:20.570475 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod85c1cb1a_6343_490c_98a3_f11a52290a73.slice/crio-94e81e500693312ac13759e34f54082033edb562b8b9e75959b6c382f08bb8a5 WatchSource:0}: Error finding container 94e81e500693312ac13759e34f54082033edb562b8b9e75959b6c382f08bb8a5: Status 404 returned error can't find the container with id 94e81e500693312ac13759e34f54082033edb562b8b9e75959b6c382f08bb8a5 Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.744746 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v6thv" event={"ID":"0c6cb779-d660-45f8-8fb2-5829d42ee030","Type":"ContainerDied","Data":"e1bcb0789d6c13f653780242f8448b2b193c29cf60ec5b474aead9d95a211a14"} Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.744793 4896 scope.go:117] "RemoveContainer" containerID="48efb299ab8ed41e49575c9749f3da5efa5277019eabe5db6fe02dcf8d920bd0" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.744809 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v6thv" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.745843 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c887bd576-rw68b" event={"ID":"85c1cb1a-6343-490c-98a3-f11a52290a73","Type":"ContainerStarted","Data":"94e81e500693312ac13759e34f54082033edb562b8b9e75959b6c382f08bb8a5"} Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.748591 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-27bbf" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.748609 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-27bbf" event={"ID":"31b0687b-5c42-4259-b96d-284c1aa487b4","Type":"ContainerDied","Data":"c8fce6024a05cf318e4dc5dafd64d062e48c552ff0be2303465123c97e71e6b8"} Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.749992 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"4a847ad6-23c1-43e2-b916-c97df45c2aff","Type":"ContainerStarted","Data":"bf35330fd57e30497c9c43ee40c3f185dc86c6a6034b62d94debd3f52ec6439f"} Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.752847 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"dfcee59f-69ac-42ac-86e8-20c08066d0c1","Type":"ContainerStarted","Data":"0a7f5155f3a549dff120ed9fc1c4b51833fc5411cab57894ab7a69ae822ebc52"} Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.755106 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-98x74" event={"ID":"229213f7-7a25-4982-a449-0122912ba05c","Type":"ContainerStarted","Data":"875182ecbd22228942b249a9304fb4f6e72ed4277c43933b1aa235342efb5295"} Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.755201 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-98x74" event={"ID":"229213f7-7a25-4982-a449-0122912ba05c","Type":"ContainerStarted","Data":"1a5369f27bfadc6a5272fb322d7d756de7870dd7fec528b00794c781fe1f94c1"} Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.763441 4896 scope.go:117] "RemoveContainer" containerID="6709c0c6bc6eb8fad3c1e2c00d5547edbd703aa0fbbbea4809220b10b020f485" Feb 18 14:58:20 crc kubenswrapper[4896]: E0218 14:58:20.763921 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-46f87" podUID="aa56560c-aab8-481c-9e3e-ac97ed9b69a5" Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.787904 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-27bbf"] Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.794826 4896 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-27bbf"] Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.798281 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-v6thv"] Feb 18 14:58:20 crc kubenswrapper[4896]: I0218 14:58:20.800682 4896 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-v6thv"] Feb 18 14:58:21 crc kubenswrapper[4896]: E0218 14:58:21.754313 4896 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 18 14:58:21 crc kubenswrapper[4896]: E0218 14:58:21.754722 4896 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bjjt6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-7s2vj_openshift-marketplace(cdc2c09e-a873-4f65-8a5f-949b4562cb10): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 18 14:58:21 crc kubenswrapper[4896]: E0218 14:58:21.756499 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-7s2vj" podUID="cdc2c09e-a873-4f65-8a5f-949b4562cb10" Feb 18 14:58:21 crc kubenswrapper[4896]: I0218 14:58:21.764211 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c887bd576-rw68b" event={"ID":"85c1cb1a-6343-490c-98a3-f11a52290a73","Type":"ContainerStarted","Data":"de9e0424956f2a0744091ce1d08afeeedd630f291f0e017c8cc1ad4e08ea4db4"} Feb 18 14:58:21 crc kubenswrapper[4896]: I0218 14:58:21.766452 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"4a847ad6-23c1-43e2-b916-c97df45c2aff","Type":"ContainerStarted","Data":"00702078018dc14dc016ecccfdb84e35551d242eec8598812df3356279f9e740"} Feb 18 14:58:21 crc kubenswrapper[4896]: I0218 14:58:21.768388 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"dfcee59f-69ac-42ac-86e8-20c08066d0c1","Type":"ContainerStarted","Data":"a7f91025ff83975b33e40e29c38fcb7eeb434b282dc429c2a82b3b1a40d9deec"} Feb 18 14:58:21 crc kubenswrapper[4896]: E0218 14:58:21.770029 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-7s2vj" podUID="cdc2c09e-a873-4f65-8a5f-949b4562cb10" Feb 18 14:58:21 crc kubenswrapper[4896]: I0218 14:58:21.935508 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c6cb779-d660-45f8-8fb2-5829d42ee030" path="/var/lib/kubelet/pods/0c6cb779-d660-45f8-8fb2-5829d42ee030/volumes" Feb 18 14:58:21 crc kubenswrapper[4896]: I0218 14:58:21.940692 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31b0687b-5c42-4259-b96d-284c1aa487b4" path="/var/lib/kubelet/pods/31b0687b-5c42-4259-b96d-284c1aa487b4/volumes" Feb 18 14:58:22 crc kubenswrapper[4896]: E0218 14:58:22.201319 4896 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 18 14:58:22 crc kubenswrapper[4896]: E0218 14:58:22.201481 4896 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8vtkk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-cmxh5_openshift-marketplace(d6bf362d-c626-4e5d-8ff4-6e847ef4eb47): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 18 14:58:22 crc kubenswrapper[4896]: E0218 14:58:22.202739 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-cmxh5" podUID="d6bf362d-c626-4e5d-8ff4-6e847ef4eb47" Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.229365 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-c6db4b444-cr7c9"] Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.231792 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-c6db4b444-cr7c9" Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.233830 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.233872 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.233915 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.234051 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.234115 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.234994 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.242603 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-c6db4b444-cr7c9"] Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.247222 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.283448 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e2590801-a8b2-442a-af31-b03b63138aa3-client-ca\") pod \"controller-manager-c6db4b444-cr7c9\" (UID: \"e2590801-a8b2-442a-af31-b03b63138aa3\") " pod="openshift-controller-manager/controller-manager-c6db4b444-cr7c9" Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.283526 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e2590801-a8b2-442a-af31-b03b63138aa3-proxy-ca-bundles\") pod \"controller-manager-c6db4b444-cr7c9\" (UID: \"e2590801-a8b2-442a-af31-b03b63138aa3\") " pod="openshift-controller-manager/controller-manager-c6db4b444-cr7c9" Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.283552 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e2590801-a8b2-442a-af31-b03b63138aa3-serving-cert\") pod \"controller-manager-c6db4b444-cr7c9\" (UID: \"e2590801-a8b2-442a-af31-b03b63138aa3\") " pod="openshift-controller-manager/controller-manager-c6db4b444-cr7c9" Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.283587 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2590801-a8b2-442a-af31-b03b63138aa3-config\") pod \"controller-manager-c6db4b444-cr7c9\" (UID: \"e2590801-a8b2-442a-af31-b03b63138aa3\") " pod="openshift-controller-manager/controller-manager-c6db4b444-cr7c9" Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.283632 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2xhl\" (UniqueName: \"kubernetes.io/projected/e2590801-a8b2-442a-af31-b03b63138aa3-kube-api-access-m2xhl\") pod \"controller-manager-c6db4b444-cr7c9\" (UID: \"e2590801-a8b2-442a-af31-b03b63138aa3\") " pod="openshift-controller-manager/controller-manager-c6db4b444-cr7c9" Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.385022 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e2590801-a8b2-442a-af31-b03b63138aa3-client-ca\") pod \"controller-manager-c6db4b444-cr7c9\" (UID: \"e2590801-a8b2-442a-af31-b03b63138aa3\") " pod="openshift-controller-manager/controller-manager-c6db4b444-cr7c9" Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.385087 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e2590801-a8b2-442a-af31-b03b63138aa3-proxy-ca-bundles\") pod \"controller-manager-c6db4b444-cr7c9\" (UID: \"e2590801-a8b2-442a-af31-b03b63138aa3\") " pod="openshift-controller-manager/controller-manager-c6db4b444-cr7c9" Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.385110 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e2590801-a8b2-442a-af31-b03b63138aa3-serving-cert\") pod \"controller-manager-c6db4b444-cr7c9\" (UID: \"e2590801-a8b2-442a-af31-b03b63138aa3\") " pod="openshift-controller-manager/controller-manager-c6db4b444-cr7c9" Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.385137 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2590801-a8b2-442a-af31-b03b63138aa3-config\") pod \"controller-manager-c6db4b444-cr7c9\" (UID: \"e2590801-a8b2-442a-af31-b03b63138aa3\") " pod="openshift-controller-manager/controller-manager-c6db4b444-cr7c9" Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.385168 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2xhl\" (UniqueName: \"kubernetes.io/projected/e2590801-a8b2-442a-af31-b03b63138aa3-kube-api-access-m2xhl\") pod \"controller-manager-c6db4b444-cr7c9\" (UID: \"e2590801-a8b2-442a-af31-b03b63138aa3\") " pod="openshift-controller-manager/controller-manager-c6db4b444-cr7c9" Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.386074 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e2590801-a8b2-442a-af31-b03b63138aa3-client-ca\") pod \"controller-manager-c6db4b444-cr7c9\" (UID: \"e2590801-a8b2-442a-af31-b03b63138aa3\") " pod="openshift-controller-manager/controller-manager-c6db4b444-cr7c9" Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.386672 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2590801-a8b2-442a-af31-b03b63138aa3-config\") pod \"controller-manager-c6db4b444-cr7c9\" (UID: \"e2590801-a8b2-442a-af31-b03b63138aa3\") " pod="openshift-controller-manager/controller-manager-c6db4b444-cr7c9" Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.387699 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e2590801-a8b2-442a-af31-b03b63138aa3-proxy-ca-bundles\") pod \"controller-manager-c6db4b444-cr7c9\" (UID: \"e2590801-a8b2-442a-af31-b03b63138aa3\") " pod="openshift-controller-manager/controller-manager-c6db4b444-cr7c9" Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.399903 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e2590801-a8b2-442a-af31-b03b63138aa3-serving-cert\") pod \"controller-manager-c6db4b444-cr7c9\" (UID: \"e2590801-a8b2-442a-af31-b03b63138aa3\") " pod="openshift-controller-manager/controller-manager-c6db4b444-cr7c9" Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.402394 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2xhl\" (UniqueName: \"kubernetes.io/projected/e2590801-a8b2-442a-af31-b03b63138aa3-kube-api-access-m2xhl\") pod \"controller-manager-c6db4b444-cr7c9\" (UID: \"e2590801-a8b2-442a-af31-b03b63138aa3\") " pod="openshift-controller-manager/controller-manager-c6db4b444-cr7c9" Feb 18 14:58:22 crc kubenswrapper[4896]: E0218 14:58:22.439003 4896 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 18 14:58:22 crc kubenswrapper[4896]: E0218 14:58:22.439150 4896 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hsqjj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-cp8c2_openshift-marketplace(96fbf244-90e4-4ffb-8dab-1ad893364032): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 18 14:58:22 crc kubenswrapper[4896]: E0218 14:58:22.440467 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-cp8c2" podUID="96fbf244-90e4-4ffb-8dab-1ad893364032" Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.555613 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-c6db4b444-cr7c9" Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.758080 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-c6db4b444-cr7c9"] Feb 18 14:58:22 crc kubenswrapper[4896]: W0218 14:58:22.764808 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode2590801_a8b2_442a_af31_b03b63138aa3.slice/crio-2663fae556f2ab06c94adf2594e4112106632266d939fa459b02970aaad80d66 WatchSource:0}: Error finding container 2663fae556f2ab06c94adf2594e4112106632266d939fa459b02970aaad80d66: Status 404 returned error can't find the container with id 2663fae556f2ab06c94adf2594e4112106632266d939fa459b02970aaad80d66 Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.775478 4896 generic.go:334] "Generic (PLEG): container finished" podID="dfcee59f-69ac-42ac-86e8-20c08066d0c1" containerID="a7f91025ff83975b33e40e29c38fcb7eeb434b282dc429c2a82b3b1a40d9deec" exitCode=0 Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.775874 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"dfcee59f-69ac-42ac-86e8-20c08066d0c1","Type":"ContainerDied","Data":"a7f91025ff83975b33e40e29c38fcb7eeb434b282dc429c2a82b3b1a40d9deec"} Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.778630 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-98x74" event={"ID":"229213f7-7a25-4982-a449-0122912ba05c","Type":"ContainerStarted","Data":"c472b6413d617344fcf26e76eccc3045ac90eceb56eae5777616f412c27366ea"} Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.780359 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-c6db4b444-cr7c9" event={"ID":"e2590801-a8b2-442a-af31-b03b63138aa3","Type":"ContainerStarted","Data":"2663fae556f2ab06c94adf2594e4112106632266d939fa459b02970aaad80d66"} Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.780583 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-c887bd576-rw68b" Feb 18 14:58:22 crc kubenswrapper[4896]: E0218 14:58:22.781032 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-cp8c2" podUID="96fbf244-90e4-4ffb-8dab-1ad893364032" Feb 18 14:58:22 crc kubenswrapper[4896]: E0218 14:58:22.781310 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-cmxh5" podUID="d6bf362d-c626-4e5d-8ff4-6e847ef4eb47" Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.786749 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-c887bd576-rw68b" Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.844534 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-c887bd576-rw68b" podStartSLOduration=12.844514678 podStartE2EDuration="12.844514678s" podCreationTimestamp="2026-02-18 14:58:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:58:22.844320342 +0000 UTC m=+201.773654555" watchObservedRunningTime="2026-02-18 14:58:22.844514678 +0000 UTC m=+201.773848901" Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.864189 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=8.864173228 podStartE2EDuration="8.864173228s" podCreationTimestamp="2026-02-18 14:58:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:58:22.860942424 +0000 UTC m=+201.790276627" watchObservedRunningTime="2026-02-18 14:58:22.864173228 +0000 UTC m=+201.793507441" Feb 18 14:58:22 crc kubenswrapper[4896]: I0218 14:58:22.877178 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-98x74" podStartSLOduration=178.877162905 podStartE2EDuration="2m58.877162905s" podCreationTimestamp="2026-02-18 14:55:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:58:22.875715423 +0000 UTC m=+201.805049636" watchObservedRunningTime="2026-02-18 14:58:22.877162905 +0000 UTC m=+201.806497118" Feb 18 14:58:23 crc kubenswrapper[4896]: I0218 14:58:23.786571 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-c6db4b444-cr7c9" event={"ID":"e2590801-a8b2-442a-af31-b03b63138aa3","Type":"ContainerStarted","Data":"2f1fde29ede1d3cd3fe1ce2d38113ba249cd5ba85e7da9b6748379d9eba796fb"} Feb 18 14:58:23 crc kubenswrapper[4896]: I0218 14:58:23.810986 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-c6db4b444-cr7c9" podStartSLOduration=13.810964056 podStartE2EDuration="13.810964056s" podCreationTimestamp="2026-02-18 14:58:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:58:23.806739064 +0000 UTC m=+202.736073277" watchObservedRunningTime="2026-02-18 14:58:23.810964056 +0000 UTC m=+202.740298269" Feb 18 14:58:24 crc kubenswrapper[4896]: I0218 14:58:24.055831 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 18 14:58:24 crc kubenswrapper[4896]: I0218 14:58:24.110786 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dfcee59f-69ac-42ac-86e8-20c08066d0c1-kube-api-access\") pod \"dfcee59f-69ac-42ac-86e8-20c08066d0c1\" (UID: \"dfcee59f-69ac-42ac-86e8-20c08066d0c1\") " Feb 18 14:58:24 crc kubenswrapper[4896]: I0218 14:58:24.110868 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dfcee59f-69ac-42ac-86e8-20c08066d0c1-kubelet-dir\") pod \"dfcee59f-69ac-42ac-86e8-20c08066d0c1\" (UID: \"dfcee59f-69ac-42ac-86e8-20c08066d0c1\") " Feb 18 14:58:24 crc kubenswrapper[4896]: I0218 14:58:24.111194 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dfcee59f-69ac-42ac-86e8-20c08066d0c1-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "dfcee59f-69ac-42ac-86e8-20c08066d0c1" (UID: "dfcee59f-69ac-42ac-86e8-20c08066d0c1"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 14:58:24 crc kubenswrapper[4896]: I0218 14:58:24.117785 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfcee59f-69ac-42ac-86e8-20c08066d0c1-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "dfcee59f-69ac-42ac-86e8-20c08066d0c1" (UID: "dfcee59f-69ac-42ac-86e8-20c08066d0c1"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:58:24 crc kubenswrapper[4896]: I0218 14:58:24.212692 4896 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dfcee59f-69ac-42ac-86e8-20c08066d0c1-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 18 14:58:24 crc kubenswrapper[4896]: I0218 14:58:24.213094 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dfcee59f-69ac-42ac-86e8-20c08066d0c1-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 18 14:58:24 crc kubenswrapper[4896]: I0218 14:58:24.795436 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"dfcee59f-69ac-42ac-86e8-20c08066d0c1","Type":"ContainerDied","Data":"0a7f5155f3a549dff120ed9fc1c4b51833fc5411cab57894ab7a69ae822ebc52"} Feb 18 14:58:24 crc kubenswrapper[4896]: I0218 14:58:24.795516 4896 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a7f5155f3a549dff120ed9fc1c4b51833fc5411cab57894ab7a69ae822ebc52" Feb 18 14:58:24 crc kubenswrapper[4896]: I0218 14:58:24.795453 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 18 14:58:24 crc kubenswrapper[4896]: I0218 14:58:24.795891 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-c6db4b444-cr7c9" Feb 18 14:58:24 crc kubenswrapper[4896]: I0218 14:58:24.801436 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-c6db4b444-cr7c9" Feb 18 14:58:34 crc kubenswrapper[4896]: I0218 14:58:34.075843 4896 patch_prober.go:28] interesting pod/machine-config-daemon-24rq5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 14:58:34 crc kubenswrapper[4896]: I0218 14:58:34.076566 4896 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 14:58:34 crc kubenswrapper[4896]: I0218 14:58:34.076643 4896 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" Feb 18 14:58:34 crc kubenswrapper[4896]: I0218 14:58:34.077630 4896 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc"} pod="openshift-machine-config-operator/machine-config-daemon-24rq5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 18 14:58:34 crc kubenswrapper[4896]: I0218 14:58:34.077852 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" containerID="cri-o://5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc" gracePeriod=600 Feb 18 14:58:34 crc kubenswrapper[4896]: I0218 14:58:34.864627 4896 generic.go:334] "Generic (PLEG): container finished" podID="8503734a-8111-4fab-beda-1f0f3b59615b" containerID="5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc" exitCode=0 Feb 18 14:58:34 crc kubenswrapper[4896]: I0218 14:58:34.864708 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" event={"ID":"8503734a-8111-4fab-beda-1f0f3b59615b","Type":"ContainerDied","Data":"5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc"} Feb 18 14:58:35 crc kubenswrapper[4896]: I0218 14:58:35.871300 4896 generic.go:334] "Generic (PLEG): container finished" podID="b8604b65-3d0c-40e3-93bb-00ff1169f8f9" containerID="aa751a5090abfb9a77dadd6cfd8c664132c3aab4d7df2440826785ed07af0c26" exitCode=0 Feb 18 14:58:35 crc kubenswrapper[4896]: I0218 14:58:35.871368 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gq2n2" event={"ID":"b8604b65-3d0c-40e3-93bb-00ff1169f8f9","Type":"ContainerDied","Data":"aa751a5090abfb9a77dadd6cfd8c664132c3aab4d7df2440826785ed07af0c26"} Feb 18 14:58:35 crc kubenswrapper[4896]: I0218 14:58:35.875102 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" event={"ID":"8503734a-8111-4fab-beda-1f0f3b59615b","Type":"ContainerStarted","Data":"aa2586e82cda0fb6f85fe0dbdbf47b3474eef31fa5931dd1eca7bed587048bb4"} Feb 18 14:58:35 crc kubenswrapper[4896]: I0218 14:58:35.876527 4896 generic.go:334] "Generic (PLEG): container finished" podID="aa56560c-aab8-481c-9e3e-ac97ed9b69a5" containerID="4c8d9c6a0e95bc6eb06861ee3753133baab9c1f1d84a35c7c6e5b1dc843dde32" exitCode=0 Feb 18 14:58:35 crc kubenswrapper[4896]: I0218 14:58:35.876582 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-46f87" event={"ID":"aa56560c-aab8-481c-9e3e-ac97ed9b69a5","Type":"ContainerDied","Data":"4c8d9c6a0e95bc6eb06861ee3753133baab9c1f1d84a35c7c6e5b1dc843dde32"} Feb 18 14:58:35 crc kubenswrapper[4896]: I0218 14:58:35.877824 4896 generic.go:334] "Generic (PLEG): container finished" podID="c93c74b9-3b90-4f6d-ab3b-8210d0704ce0" containerID="a125354f7050d241c6921b5af4109c06b9928effbf19d77d79f829a9f8570eba" exitCode=0 Feb 18 14:58:35 crc kubenswrapper[4896]: I0218 14:58:35.877839 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4s8kn" event={"ID":"c93c74b9-3b90-4f6d-ab3b-8210d0704ce0","Type":"ContainerDied","Data":"a125354f7050d241c6921b5af4109c06b9928effbf19d77d79f829a9f8570eba"} Feb 18 14:58:35 crc kubenswrapper[4896]: I0218 14:58:35.879101 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t6sfb" event={"ID":"eddbcdff-a387-4f95-9277-5850b6485036","Type":"ContainerStarted","Data":"d943c2184f788626ab04198b6da0477b59006414dee16d46906e39d9fb3600f0"} Feb 18 14:58:36 crc kubenswrapper[4896]: I0218 14:58:36.886466 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cp8c2" event={"ID":"96fbf244-90e4-4ffb-8dab-1ad893364032","Type":"ContainerStarted","Data":"41bc3f54497918a30d8c5eb72b232b41fb29216fe9fa56db8a9eb6c4394a1317"} Feb 18 14:58:36 crc kubenswrapper[4896]: I0218 14:58:36.888594 4896 generic.go:334] "Generic (PLEG): container finished" podID="eddbcdff-a387-4f95-9277-5850b6485036" containerID="d943c2184f788626ab04198b6da0477b59006414dee16d46906e39d9fb3600f0" exitCode=0 Feb 18 14:58:36 crc kubenswrapper[4896]: I0218 14:58:36.888651 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t6sfb" event={"ID":"eddbcdff-a387-4f95-9277-5850b6485036","Type":"ContainerDied","Data":"d943c2184f788626ab04198b6da0477b59006414dee16d46906e39d9fb3600f0"} Feb 18 14:58:36 crc kubenswrapper[4896]: I0218 14:58:36.897907 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wzwx5" event={"ID":"f48b5c5e-4ec2-4bca-bd26-02b837c029ea","Type":"ContainerStarted","Data":"fc8ab050528b7790a11f300df0018b5123625f59a8f21c6938a2d5baf6b3afcb"} Feb 18 14:58:37 crc kubenswrapper[4896]: I0218 14:58:37.910687 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gq2n2" event={"ID":"b8604b65-3d0c-40e3-93bb-00ff1169f8f9","Type":"ContainerStarted","Data":"4a7340c954072b2dbf2cb422d0aa88eed139f6a041257cbe8fb727ff319f0cb1"} Feb 18 14:58:37 crc kubenswrapper[4896]: I0218 14:58:37.914444 4896 generic.go:334] "Generic (PLEG): container finished" podID="f48b5c5e-4ec2-4bca-bd26-02b837c029ea" containerID="fc8ab050528b7790a11f300df0018b5123625f59a8f21c6938a2d5baf6b3afcb" exitCode=0 Feb 18 14:58:37 crc kubenswrapper[4896]: I0218 14:58:37.914521 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wzwx5" event={"ID":"f48b5c5e-4ec2-4bca-bd26-02b837c029ea","Type":"ContainerDied","Data":"fc8ab050528b7790a11f300df0018b5123625f59a8f21c6938a2d5baf6b3afcb"} Feb 18 14:58:37 crc kubenswrapper[4896]: I0218 14:58:37.916402 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-46f87" event={"ID":"aa56560c-aab8-481c-9e3e-ac97ed9b69a5","Type":"ContainerStarted","Data":"8f90a1e466747e8c2c4a90531097356a6c13422797c456afbab5da533704f60d"} Feb 18 14:58:37 crc kubenswrapper[4896]: I0218 14:58:37.917921 4896 generic.go:334] "Generic (PLEG): container finished" podID="96fbf244-90e4-4ffb-8dab-1ad893364032" containerID="41bc3f54497918a30d8c5eb72b232b41fb29216fe9fa56db8a9eb6c4394a1317" exitCode=0 Feb 18 14:58:37 crc kubenswrapper[4896]: I0218 14:58:37.917954 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cp8c2" event={"ID":"96fbf244-90e4-4ffb-8dab-1ad893364032","Type":"ContainerDied","Data":"41bc3f54497918a30d8c5eb72b232b41fb29216fe9fa56db8a9eb6c4394a1317"} Feb 18 14:58:37 crc kubenswrapper[4896]: I0218 14:58:37.920864 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4s8kn" event={"ID":"c93c74b9-3b90-4f6d-ab3b-8210d0704ce0","Type":"ContainerStarted","Data":"a115e2c3525f35515f16d0eb755b14ed75da309a8480c53586ace1ddb3813d2b"} Feb 18 14:58:37 crc kubenswrapper[4896]: I0218 14:58:37.931312 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gq2n2" podStartSLOduration=3.094348456 podStartE2EDuration="1m3.931297038s" podCreationTimestamp="2026-02-18 14:57:34 +0000 UTC" firstStartedPulling="2026-02-18 14:57:36.434699084 +0000 UTC m=+155.364033297" lastFinishedPulling="2026-02-18 14:58:37.271647666 +0000 UTC m=+216.200981879" observedRunningTime="2026-02-18 14:58:37.930409722 +0000 UTC m=+216.859743935" watchObservedRunningTime="2026-02-18 14:58:37.931297038 +0000 UTC m=+216.860631241" Feb 18 14:58:37 crc kubenswrapper[4896]: I0218 14:58:37.949253 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-46f87" podStartSLOduration=3.74792471 podStartE2EDuration="1m7.949224048s" podCreationTimestamp="2026-02-18 14:57:30 +0000 UTC" firstStartedPulling="2026-02-18 14:57:33.349394589 +0000 UTC m=+152.278728802" lastFinishedPulling="2026-02-18 14:58:37.550693927 +0000 UTC m=+216.480028140" observedRunningTime="2026-02-18 14:58:37.946948382 +0000 UTC m=+216.876282595" watchObservedRunningTime="2026-02-18 14:58:37.949224048 +0000 UTC m=+216.878558261" Feb 18 14:58:37 crc kubenswrapper[4896]: I0218 14:58:37.977995 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4s8kn" podStartSLOduration=1.783878708 podStartE2EDuration="1m4.977974143s" podCreationTimestamp="2026-02-18 14:57:33 +0000 UTC" firstStartedPulling="2026-02-18 14:57:34.378295352 +0000 UTC m=+153.307629565" lastFinishedPulling="2026-02-18 14:58:37.572390787 +0000 UTC m=+216.501725000" observedRunningTime="2026-02-18 14:58:37.977045046 +0000 UTC m=+216.906379259" watchObservedRunningTime="2026-02-18 14:58:37.977974143 +0000 UTC m=+216.907308356" Feb 18 14:58:38 crc kubenswrapper[4896]: I0218 14:58:38.929965 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t6sfb" event={"ID":"eddbcdff-a387-4f95-9277-5850b6485036","Type":"ContainerStarted","Data":"6a9b0468659ba66ab625173e6087c0b54094b68f2dc654936d5502396abaa27c"} Feb 18 14:58:38 crc kubenswrapper[4896]: I0218 14:58:38.932612 4896 generic.go:334] "Generic (PLEG): container finished" podID="cdc2c09e-a873-4f65-8a5f-949b4562cb10" containerID="bff0a899580e45f76dbbd0c22366ad2c6851ae2debe469224ea73c8f33c86f92" exitCode=0 Feb 18 14:58:38 crc kubenswrapper[4896]: I0218 14:58:38.932693 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7s2vj" event={"ID":"cdc2c09e-a873-4f65-8a5f-949b4562cb10","Type":"ContainerDied","Data":"bff0a899580e45f76dbbd0c22366ad2c6851ae2debe469224ea73c8f33c86f92"} Feb 18 14:58:38 crc kubenswrapper[4896]: I0218 14:58:38.971484 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-t6sfb" podStartSLOduration=3.384333455 podStartE2EDuration="1m4.971461287s" podCreationTimestamp="2026-02-18 14:57:34 +0000 UTC" firstStartedPulling="2026-02-18 14:57:36.460639567 +0000 UTC m=+155.389973780" lastFinishedPulling="2026-02-18 14:58:38.047767399 +0000 UTC m=+216.977101612" observedRunningTime="2026-02-18 14:58:38.9505603 +0000 UTC m=+217.879894513" watchObservedRunningTime="2026-02-18 14:58:38.971461287 +0000 UTC m=+217.900795510" Feb 18 14:58:40 crc kubenswrapper[4896]: I0218 14:58:40.949172 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cmxh5" event={"ID":"d6bf362d-c626-4e5d-8ff4-6e847ef4eb47","Type":"ContainerStarted","Data":"87bc842934ed7f6a0934b6c5bc66530fc2f47ebc7203927496453fcd790d2ab4"} Feb 18 14:58:40 crc kubenswrapper[4896]: I0218 14:58:40.963557 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cp8c2" event={"ID":"96fbf244-90e4-4ffb-8dab-1ad893364032","Type":"ContainerStarted","Data":"cfba7a7ca5cc789db0990395717770a98b6c2de038a044c1e687035aad716e5f"} Feb 18 14:58:41 crc kubenswrapper[4896]: I0218 14:58:41.170746 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-46f87" Feb 18 14:58:41 crc kubenswrapper[4896]: I0218 14:58:41.171050 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-46f87" Feb 18 14:58:41 crc kubenswrapper[4896]: I0218 14:58:41.962254 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-46f87" Feb 18 14:58:41 crc kubenswrapper[4896]: I0218 14:58:41.970029 4896 generic.go:334] "Generic (PLEG): container finished" podID="d6bf362d-c626-4e5d-8ff4-6e847ef4eb47" containerID="87bc842934ed7f6a0934b6c5bc66530fc2f47ebc7203927496453fcd790d2ab4" exitCode=0 Feb 18 14:58:41 crc kubenswrapper[4896]: I0218 14:58:41.970070 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cmxh5" event={"ID":"d6bf362d-c626-4e5d-8ff4-6e847ef4eb47","Type":"ContainerDied","Data":"87bc842934ed7f6a0934b6c5bc66530fc2f47ebc7203927496453fcd790d2ab4"} Feb 18 14:58:42 crc kubenswrapper[4896]: I0218 14:58:42.000144 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cp8c2" podStartSLOduration=4.000312969 podStartE2EDuration="1m11.000110757s" podCreationTimestamp="2026-02-18 14:57:31 +0000 UTC" firstStartedPulling="2026-02-18 14:57:33.333761425 +0000 UTC m=+152.263095638" lastFinishedPulling="2026-02-18 14:58:40.333559203 +0000 UTC m=+219.262893426" observedRunningTime="2026-02-18 14:58:41.997293395 +0000 UTC m=+220.926627608" watchObservedRunningTime="2026-02-18 14:58:42.000110757 +0000 UTC m=+220.929444980" Feb 18 14:58:42 crc kubenswrapper[4896]: I0218 14:58:42.977526 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wzwx5" event={"ID":"f48b5c5e-4ec2-4bca-bd26-02b837c029ea","Type":"ContainerStarted","Data":"f667d245c6fbeb79de087de71855903712f2a50648b8e588f04b173ed6451daf"} Feb 18 14:58:43 crc kubenswrapper[4896]: I0218 14:58:43.009691 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wzwx5" podStartSLOduration=3.728342624 podStartE2EDuration="1m12.009670938s" podCreationTimestamp="2026-02-18 14:57:31 +0000 UTC" firstStartedPulling="2026-02-18 14:57:33.344300411 +0000 UTC m=+152.273634624" lastFinishedPulling="2026-02-18 14:58:41.625628725 +0000 UTC m=+220.554962938" observedRunningTime="2026-02-18 14:58:43.008745731 +0000 UTC m=+221.938079944" watchObservedRunningTime="2026-02-18 14:58:43.009670938 +0000 UTC m=+221.939005151" Feb 18 14:58:43 crc kubenswrapper[4896]: I0218 14:58:43.028500 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-46f87" Feb 18 14:58:43 crc kubenswrapper[4896]: I0218 14:58:43.561560 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4s8kn" Feb 18 14:58:43 crc kubenswrapper[4896]: I0218 14:58:43.561823 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4s8kn" Feb 18 14:58:43 crc kubenswrapper[4896]: I0218 14:58:43.608569 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4s8kn" Feb 18 14:58:43 crc kubenswrapper[4896]: I0218 14:58:43.985123 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cmxh5" event={"ID":"d6bf362d-c626-4e5d-8ff4-6e847ef4eb47","Type":"ContainerStarted","Data":"0856168e13497c09e9b7ac5a43826f510095d802c443a5df89e2e149037e3508"} Feb 18 14:58:43 crc kubenswrapper[4896]: I0218 14:58:43.986818 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7s2vj" event={"ID":"cdc2c09e-a873-4f65-8a5f-949b4562cb10","Type":"ContainerStarted","Data":"09cceded19dc3ecf87f307ea5dc21fe28e3836f84630322a8cdce3a11ed12dbb"} Feb 18 14:58:44 crc kubenswrapper[4896]: I0218 14:58:44.005952 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-cmxh5" podStartSLOduration=2.612053398 podStartE2EDuration="1m12.005932722s" podCreationTimestamp="2026-02-18 14:57:32 +0000 UTC" firstStartedPulling="2026-02-18 14:57:34.374757539 +0000 UTC m=+153.304091752" lastFinishedPulling="2026-02-18 14:58:43.768636863 +0000 UTC m=+222.697971076" observedRunningTime="2026-02-18 14:58:44.001470093 +0000 UTC m=+222.930804306" watchObservedRunningTime="2026-02-18 14:58:44.005932722 +0000 UTC m=+222.935266935" Feb 18 14:58:44 crc kubenswrapper[4896]: I0218 14:58:44.022781 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7s2vj" podStartSLOduration=3.537853035 podStartE2EDuration="1m13.022761811s" podCreationTimestamp="2026-02-18 14:57:31 +0000 UTC" firstStartedPulling="2026-02-18 14:57:33.339281846 +0000 UTC m=+152.268616059" lastFinishedPulling="2026-02-18 14:58:42.824190622 +0000 UTC m=+221.753524835" observedRunningTime="2026-02-18 14:58:44.021522765 +0000 UTC m=+222.950856978" watchObservedRunningTime="2026-02-18 14:58:44.022761811 +0000 UTC m=+222.952096024" Feb 18 14:58:44 crc kubenswrapper[4896]: I0218 14:58:44.050743 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4s8kn" Feb 18 14:58:44 crc kubenswrapper[4896]: I0218 14:58:44.582080 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gq2n2" Feb 18 14:58:44 crc kubenswrapper[4896]: I0218 14:58:44.582460 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gq2n2" Feb 18 14:58:48 crc kubenswrapper[4896]: I0218 14:58:48.012306 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-t6sfb" Feb 18 14:58:48 crc kubenswrapper[4896]: I0218 14:58:48.042789 4896 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-shvc6 container/oauth-openshift namespace/openshift-authentication: Liveness probe status=failure output="Get \"https://10.217.0.29:6443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 18 14:58:48 crc kubenswrapper[4896]: I0218 14:58:48.043224 4896 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" podUID="76cb1e0e-4321-4450-9fcb-b505a7a1d969" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.29:6443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 18 14:58:48 crc kubenswrapper[4896]: I0218 14:58:48.043533 4896 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-shvc6 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.29:6443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 18 14:58:48 crc kubenswrapper[4896]: I0218 14:58:48.043564 4896 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" podUID="76cb1e0e-4321-4450-9fcb-b505a7a1d969" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.29:6443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 18 14:58:48 crc kubenswrapper[4896]: I0218 14:58:48.050437 4896 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gq2n2" podUID="b8604b65-3d0c-40e3-93bb-00ff1169f8f9" containerName="registry-server" probeResult="failure" output=< Feb 18 14:58:48 crc kubenswrapper[4896]: timeout: failed to connect service ":50051" within 1s Feb 18 14:58:48 crc kubenswrapper[4896]: > Feb 18 14:58:48 crc kubenswrapper[4896]: I0218 14:58:48.090515 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-t6sfb" Feb 18 14:58:48 crc kubenswrapper[4896]: I0218 14:58:48.106446 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4s8kn"] Feb 18 14:58:49 crc kubenswrapper[4896]: I0218 14:58:49.134591 4896 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-t6sfb" podUID="eddbcdff-a387-4f95-9277-5850b6485036" containerName="registry-server" probeResult="failure" output=< Feb 18 14:58:49 crc kubenswrapper[4896]: timeout: failed to connect service ":50051" within 1s Feb 18 14:58:49 crc kubenswrapper[4896]: > Feb 18 14:58:50 crc kubenswrapper[4896]: I0218 14:58:50.067710 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4s8kn" podUID="c93c74b9-3b90-4f6d-ab3b-8210d0704ce0" containerName="registry-server" containerID="cri-o://a115e2c3525f35515f16d0eb755b14ed75da309a8480c53586ace1ddb3813d2b" gracePeriod=2 Feb 18 14:58:51 crc kubenswrapper[4896]: I0218 14:58:51.077219 4896 generic.go:334] "Generic (PLEG): container finished" podID="c93c74b9-3b90-4f6d-ab3b-8210d0704ce0" containerID="a115e2c3525f35515f16d0eb755b14ed75da309a8480c53586ace1ddb3813d2b" exitCode=0 Feb 18 14:58:51 crc kubenswrapper[4896]: I0218 14:58:51.077288 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4s8kn" event={"ID":"c93c74b9-3b90-4f6d-ab3b-8210d0704ce0","Type":"ContainerDied","Data":"a115e2c3525f35515f16d0eb755b14ed75da309a8480c53586ace1ddb3813d2b"} Feb 18 14:58:51 crc kubenswrapper[4896]: I0218 14:58:51.390822 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cp8c2" Feb 18 14:58:51 crc kubenswrapper[4896]: I0218 14:58:51.391341 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cp8c2" Feb 18 14:58:51 crc kubenswrapper[4896]: I0218 14:58:51.427475 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4s8kn" Feb 18 14:58:51 crc kubenswrapper[4896]: I0218 14:58:51.449483 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cp8c2" Feb 18 14:58:51 crc kubenswrapper[4896]: I0218 14:58:51.500530 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c93c74b9-3b90-4f6d-ab3b-8210d0704ce0-utilities\") pod \"c93c74b9-3b90-4f6d-ab3b-8210d0704ce0\" (UID: \"c93c74b9-3b90-4f6d-ab3b-8210d0704ce0\") " Feb 18 14:58:51 crc kubenswrapper[4896]: I0218 14:58:51.500630 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rp6r5\" (UniqueName: \"kubernetes.io/projected/c93c74b9-3b90-4f6d-ab3b-8210d0704ce0-kube-api-access-rp6r5\") pod \"c93c74b9-3b90-4f6d-ab3b-8210d0704ce0\" (UID: \"c93c74b9-3b90-4f6d-ab3b-8210d0704ce0\") " Feb 18 14:58:51 crc kubenswrapper[4896]: I0218 14:58:51.500669 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c93c74b9-3b90-4f6d-ab3b-8210d0704ce0-catalog-content\") pod \"c93c74b9-3b90-4f6d-ab3b-8210d0704ce0\" (UID: \"c93c74b9-3b90-4f6d-ab3b-8210d0704ce0\") " Feb 18 14:58:51 crc kubenswrapper[4896]: I0218 14:58:51.501300 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c93c74b9-3b90-4f6d-ab3b-8210d0704ce0-utilities" (OuterVolumeSpecName: "utilities") pod "c93c74b9-3b90-4f6d-ab3b-8210d0704ce0" (UID: "c93c74b9-3b90-4f6d-ab3b-8210d0704ce0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 14:58:51 crc kubenswrapper[4896]: I0218 14:58:51.508051 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c93c74b9-3b90-4f6d-ab3b-8210d0704ce0-kube-api-access-rp6r5" (OuterVolumeSpecName: "kube-api-access-rp6r5") pod "c93c74b9-3b90-4f6d-ab3b-8210d0704ce0" (UID: "c93c74b9-3b90-4f6d-ab3b-8210d0704ce0"). InnerVolumeSpecName "kube-api-access-rp6r5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:58:51 crc kubenswrapper[4896]: I0218 14:58:51.523134 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c93c74b9-3b90-4f6d-ab3b-8210d0704ce0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c93c74b9-3b90-4f6d-ab3b-8210d0704ce0" (UID: "c93c74b9-3b90-4f6d-ab3b-8210d0704ce0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 14:58:51 crc kubenswrapper[4896]: I0218 14:58:51.601710 4896 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c93c74b9-3b90-4f6d-ab3b-8210d0704ce0-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 14:58:51 crc kubenswrapper[4896]: I0218 14:58:51.601756 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rp6r5\" (UniqueName: \"kubernetes.io/projected/c93c74b9-3b90-4f6d-ab3b-8210d0704ce0-kube-api-access-rp6r5\") on node \"crc\" DevicePath \"\"" Feb 18 14:58:51 crc kubenswrapper[4896]: I0218 14:58:51.601772 4896 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c93c74b9-3b90-4f6d-ab3b-8210d0704ce0-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 14:58:51 crc kubenswrapper[4896]: I0218 14:58:51.684607 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7s2vj" Feb 18 14:58:51 crc kubenswrapper[4896]: I0218 14:58:51.684654 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7s2vj" Feb 18 14:58:51 crc kubenswrapper[4896]: I0218 14:58:51.726672 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7s2vj" Feb 18 14:58:51 crc kubenswrapper[4896]: I0218 14:58:51.783789 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wzwx5" Feb 18 14:58:51 crc kubenswrapper[4896]: I0218 14:58:51.783841 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wzwx5" Feb 18 14:58:51 crc kubenswrapper[4896]: I0218 14:58:51.824086 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wzwx5" Feb 18 14:58:52 crc kubenswrapper[4896]: I0218 14:58:52.086037 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4s8kn" event={"ID":"c93c74b9-3b90-4f6d-ab3b-8210d0704ce0","Type":"ContainerDied","Data":"b17b44a752c005757e232bc39e4792bcee43fb6a109124a55311dc9357079bac"} Feb 18 14:58:52 crc kubenswrapper[4896]: I0218 14:58:52.086088 4896 scope.go:117] "RemoveContainer" containerID="a115e2c3525f35515f16d0eb755b14ed75da309a8480c53586ace1ddb3813d2b" Feb 18 14:58:52 crc kubenswrapper[4896]: I0218 14:58:52.086148 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4s8kn" Feb 18 14:58:52 crc kubenswrapper[4896]: I0218 14:58:52.104220 4896 scope.go:117] "RemoveContainer" containerID="a125354f7050d241c6921b5af4109c06b9928effbf19d77d79f829a9f8570eba" Feb 18 14:58:52 crc kubenswrapper[4896]: I0218 14:58:52.106406 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4s8kn"] Feb 18 14:58:52 crc kubenswrapper[4896]: I0218 14:58:52.110796 4896 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4s8kn"] Feb 18 14:58:52 crc kubenswrapper[4896]: I0218 14:58:52.130173 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wzwx5" Feb 18 14:58:52 crc kubenswrapper[4896]: I0218 14:58:52.131894 4896 scope.go:117] "RemoveContainer" containerID="847bb87210071ac7eb84aae032f534274e27b06285ddd17a55e3e381c7e7a989" Feb 18 14:58:52 crc kubenswrapper[4896]: I0218 14:58:52.134828 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cp8c2" Feb 18 14:58:52 crc kubenswrapper[4896]: I0218 14:58:52.138722 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7s2vj" Feb 18 14:58:53 crc kubenswrapper[4896]: I0218 14:58:53.375630 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-cmxh5" Feb 18 14:58:53 crc kubenswrapper[4896]: I0218 14:58:53.376864 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-cmxh5" Feb 18 14:58:53 crc kubenswrapper[4896]: I0218 14:58:53.415266 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-cmxh5" Feb 18 14:58:53 crc kubenswrapper[4896]: I0218 14:58:53.821890 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wzwx5"] Feb 18 14:58:53 crc kubenswrapper[4896]: I0218 14:58:53.934128 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c93c74b9-3b90-4f6d-ab3b-8210d0704ce0" path="/var/lib/kubelet/pods/c93c74b9-3b90-4f6d-ab3b-8210d0704ce0/volumes" Feb 18 14:58:54 crc kubenswrapper[4896]: I0218 14:58:54.096859 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wzwx5" podUID="f48b5c5e-4ec2-4bca-bd26-02b837c029ea" containerName="registry-server" containerID="cri-o://f667d245c6fbeb79de087de71855903712f2a50648b8e588f04b173ed6451daf" gracePeriod=2 Feb 18 14:58:54 crc kubenswrapper[4896]: I0218 14:58:54.144744 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-cmxh5" Feb 18 14:58:54 crc kubenswrapper[4896]: I0218 14:58:54.178291 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-shvc6"] Feb 18 14:58:54 crc kubenswrapper[4896]: I0218 14:58:54.418624 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7s2vj"] Feb 18 14:58:54 crc kubenswrapper[4896]: I0218 14:58:54.418829 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7s2vj" podUID="cdc2c09e-a873-4f65-8a5f-949b4562cb10" containerName="registry-server" containerID="cri-o://09cceded19dc3ecf87f307ea5dc21fe28e3836f84630322a8cdce3a11ed12dbb" gracePeriod=2 Feb 18 14:58:54 crc kubenswrapper[4896]: I0218 14:58:54.618746 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gq2n2" Feb 18 14:58:54 crc kubenswrapper[4896]: I0218 14:58:54.659427 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gq2n2" Feb 18 14:58:54 crc kubenswrapper[4896]: I0218 14:58:54.881375 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7s2vj" Feb 18 14:58:54 crc kubenswrapper[4896]: I0218 14:58:54.944259 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdc2c09e-a873-4f65-8a5f-949b4562cb10-catalog-content\") pod \"cdc2c09e-a873-4f65-8a5f-949b4562cb10\" (UID: \"cdc2c09e-a873-4f65-8a5f-949b4562cb10\") " Feb 18 14:58:54 crc kubenswrapper[4896]: I0218 14:58:54.944375 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjjt6\" (UniqueName: \"kubernetes.io/projected/cdc2c09e-a873-4f65-8a5f-949b4562cb10-kube-api-access-bjjt6\") pod \"cdc2c09e-a873-4f65-8a5f-949b4562cb10\" (UID: \"cdc2c09e-a873-4f65-8a5f-949b4562cb10\") " Feb 18 14:58:54 crc kubenswrapper[4896]: I0218 14:58:54.944424 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdc2c09e-a873-4f65-8a5f-949b4562cb10-utilities\") pod \"cdc2c09e-a873-4f65-8a5f-949b4562cb10\" (UID: \"cdc2c09e-a873-4f65-8a5f-949b4562cb10\") " Feb 18 14:58:54 crc kubenswrapper[4896]: I0218 14:58:54.945458 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cdc2c09e-a873-4f65-8a5f-949b4562cb10-utilities" (OuterVolumeSpecName: "utilities") pod "cdc2c09e-a873-4f65-8a5f-949b4562cb10" (UID: "cdc2c09e-a873-4f65-8a5f-949b4562cb10"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 14:58:54 crc kubenswrapper[4896]: I0218 14:58:54.949627 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdc2c09e-a873-4f65-8a5f-949b4562cb10-kube-api-access-bjjt6" (OuterVolumeSpecName: "kube-api-access-bjjt6") pod "cdc2c09e-a873-4f65-8a5f-949b4562cb10" (UID: "cdc2c09e-a873-4f65-8a5f-949b4562cb10"). InnerVolumeSpecName "kube-api-access-bjjt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.002167 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cdc2c09e-a873-4f65-8a5f-949b4562cb10-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cdc2c09e-a873-4f65-8a5f-949b4562cb10" (UID: "cdc2c09e-a873-4f65-8a5f-949b4562cb10"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.045477 4896 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cdc2c09e-a873-4f65-8a5f-949b4562cb10-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.045515 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjjt6\" (UniqueName: \"kubernetes.io/projected/cdc2c09e-a873-4f65-8a5f-949b4562cb10-kube-api-access-bjjt6\") on node \"crc\" DevicePath \"\"" Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.045528 4896 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cdc2c09e-a873-4f65-8a5f-949b4562cb10-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.055870 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-t6sfb" Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.096653 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-t6sfb" Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.114909 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wzwx5" Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.115012 4896 generic.go:334] "Generic (PLEG): container finished" podID="f48b5c5e-4ec2-4bca-bd26-02b837c029ea" containerID="f667d245c6fbeb79de087de71855903712f2a50648b8e588f04b173ed6451daf" exitCode=0 Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.115077 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wzwx5" event={"ID":"f48b5c5e-4ec2-4bca-bd26-02b837c029ea","Type":"ContainerDied","Data":"f667d245c6fbeb79de087de71855903712f2a50648b8e588f04b173ed6451daf"} Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.115107 4896 scope.go:117] "RemoveContainer" containerID="f667d245c6fbeb79de087de71855903712f2a50648b8e588f04b173ed6451daf" Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.117335 4896 generic.go:334] "Generic (PLEG): container finished" podID="cdc2c09e-a873-4f65-8a5f-949b4562cb10" containerID="09cceded19dc3ecf87f307ea5dc21fe28e3836f84630322a8cdce3a11ed12dbb" exitCode=0 Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.117407 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7s2vj" event={"ID":"cdc2c09e-a873-4f65-8a5f-949b4562cb10","Type":"ContainerDied","Data":"09cceded19dc3ecf87f307ea5dc21fe28e3836f84630322a8cdce3a11ed12dbb"} Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.117446 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7s2vj" event={"ID":"cdc2c09e-a873-4f65-8a5f-949b4562cb10","Type":"ContainerDied","Data":"3034d6dfd58bd6011dbc42c2834cf838b0facec47301f3048de9931fd7a695bb"} Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.117457 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7s2vj" Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.130956 4896 scope.go:117] "RemoveContainer" containerID="fc8ab050528b7790a11f300df0018b5123625f59a8f21c6938a2d5baf6b3afcb" Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.145992 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f48b5c5e-4ec2-4bca-bd26-02b837c029ea-catalog-content\") pod \"f48b5c5e-4ec2-4bca-bd26-02b837c029ea\" (UID: \"f48b5c5e-4ec2-4bca-bd26-02b837c029ea\") " Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.146073 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrtfz\" (UniqueName: \"kubernetes.io/projected/f48b5c5e-4ec2-4bca-bd26-02b837c029ea-kube-api-access-jrtfz\") pod \"f48b5c5e-4ec2-4bca-bd26-02b837c029ea\" (UID: \"f48b5c5e-4ec2-4bca-bd26-02b837c029ea\") " Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.146120 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f48b5c5e-4ec2-4bca-bd26-02b837c029ea-utilities\") pod \"f48b5c5e-4ec2-4bca-bd26-02b837c029ea\" (UID: \"f48b5c5e-4ec2-4bca-bd26-02b837c029ea\") " Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.148098 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f48b5c5e-4ec2-4bca-bd26-02b837c029ea-utilities" (OuterVolumeSpecName: "utilities") pod "f48b5c5e-4ec2-4bca-bd26-02b837c029ea" (UID: "f48b5c5e-4ec2-4bca-bd26-02b837c029ea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.152740 4896 scope.go:117] "RemoveContainer" containerID="87db43e7f5035089862903966e8817b18224516d05c374ea3d257cd62f64b3c9" Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.155132 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7s2vj"] Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.158366 4896 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7s2vj"] Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.172890 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f48b5c5e-4ec2-4bca-bd26-02b837c029ea-kube-api-access-jrtfz" (OuterVolumeSpecName: "kube-api-access-jrtfz") pod "f48b5c5e-4ec2-4bca-bd26-02b837c029ea" (UID: "f48b5c5e-4ec2-4bca-bd26-02b837c029ea"). InnerVolumeSpecName "kube-api-access-jrtfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.179104 4896 scope.go:117] "RemoveContainer" containerID="09cceded19dc3ecf87f307ea5dc21fe28e3836f84630322a8cdce3a11ed12dbb" Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.191773 4896 scope.go:117] "RemoveContainer" containerID="bff0a899580e45f76dbbd0c22366ad2c6851ae2debe469224ea73c8f33c86f92" Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.199830 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f48b5c5e-4ec2-4bca-bd26-02b837c029ea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f48b5c5e-4ec2-4bca-bd26-02b837c029ea" (UID: "f48b5c5e-4ec2-4bca-bd26-02b837c029ea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.215581 4896 scope.go:117] "RemoveContainer" containerID="f2f7bd8ffa24dec4c9003fe1d734877149cf693636e2be84537e9fc3a1331be9" Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.247083 4896 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f48b5c5e-4ec2-4bca-bd26-02b837c029ea-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.247128 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrtfz\" (UniqueName: \"kubernetes.io/projected/f48b5c5e-4ec2-4bca-bd26-02b837c029ea-kube-api-access-jrtfz\") on node \"crc\" DevicePath \"\"" Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.247140 4896 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f48b5c5e-4ec2-4bca-bd26-02b837c029ea-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.251180 4896 scope.go:117] "RemoveContainer" containerID="09cceded19dc3ecf87f307ea5dc21fe28e3836f84630322a8cdce3a11ed12dbb" Feb 18 14:58:55 crc kubenswrapper[4896]: E0218 14:58:55.251763 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09cceded19dc3ecf87f307ea5dc21fe28e3836f84630322a8cdce3a11ed12dbb\": container with ID starting with 09cceded19dc3ecf87f307ea5dc21fe28e3836f84630322a8cdce3a11ed12dbb not found: ID does not exist" containerID="09cceded19dc3ecf87f307ea5dc21fe28e3836f84630322a8cdce3a11ed12dbb" Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.251812 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09cceded19dc3ecf87f307ea5dc21fe28e3836f84630322a8cdce3a11ed12dbb"} err="failed to get container status \"09cceded19dc3ecf87f307ea5dc21fe28e3836f84630322a8cdce3a11ed12dbb\": rpc error: code = NotFound desc = could not find container \"09cceded19dc3ecf87f307ea5dc21fe28e3836f84630322a8cdce3a11ed12dbb\": container with ID starting with 09cceded19dc3ecf87f307ea5dc21fe28e3836f84630322a8cdce3a11ed12dbb not found: ID does not exist" Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.251851 4896 scope.go:117] "RemoveContainer" containerID="bff0a899580e45f76dbbd0c22366ad2c6851ae2debe469224ea73c8f33c86f92" Feb 18 14:58:55 crc kubenswrapper[4896]: E0218 14:58:55.252309 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bff0a899580e45f76dbbd0c22366ad2c6851ae2debe469224ea73c8f33c86f92\": container with ID starting with bff0a899580e45f76dbbd0c22366ad2c6851ae2debe469224ea73c8f33c86f92 not found: ID does not exist" containerID="bff0a899580e45f76dbbd0c22366ad2c6851ae2debe469224ea73c8f33c86f92" Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.252350 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bff0a899580e45f76dbbd0c22366ad2c6851ae2debe469224ea73c8f33c86f92"} err="failed to get container status \"bff0a899580e45f76dbbd0c22366ad2c6851ae2debe469224ea73c8f33c86f92\": rpc error: code = NotFound desc = could not find container \"bff0a899580e45f76dbbd0c22366ad2c6851ae2debe469224ea73c8f33c86f92\": container with ID starting with bff0a899580e45f76dbbd0c22366ad2c6851ae2debe469224ea73c8f33c86f92 not found: ID does not exist" Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.252377 4896 scope.go:117] "RemoveContainer" containerID="f2f7bd8ffa24dec4c9003fe1d734877149cf693636e2be84537e9fc3a1331be9" Feb 18 14:58:55 crc kubenswrapper[4896]: E0218 14:58:55.252636 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2f7bd8ffa24dec4c9003fe1d734877149cf693636e2be84537e9fc3a1331be9\": container with ID starting with f2f7bd8ffa24dec4c9003fe1d734877149cf693636e2be84537e9fc3a1331be9 not found: ID does not exist" containerID="f2f7bd8ffa24dec4c9003fe1d734877149cf693636e2be84537e9fc3a1331be9" Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.252667 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2f7bd8ffa24dec4c9003fe1d734877149cf693636e2be84537e9fc3a1331be9"} err="failed to get container status \"f2f7bd8ffa24dec4c9003fe1d734877149cf693636e2be84537e9fc3a1331be9\": rpc error: code = NotFound desc = could not find container \"f2f7bd8ffa24dec4c9003fe1d734877149cf693636e2be84537e9fc3a1331be9\": container with ID starting with f2f7bd8ffa24dec4c9003fe1d734877149cf693636e2be84537e9fc3a1331be9 not found: ID does not exist" Feb 18 14:58:55 crc kubenswrapper[4896]: I0218 14:58:55.934452 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdc2c09e-a873-4f65-8a5f-949b4562cb10" path="/var/lib/kubelet/pods/cdc2c09e-a873-4f65-8a5f-949b4562cb10/volumes" Feb 18 14:58:56 crc kubenswrapper[4896]: I0218 14:58:56.125073 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wzwx5" event={"ID":"f48b5c5e-4ec2-4bca-bd26-02b837c029ea","Type":"ContainerDied","Data":"a1c7b8fcec9fe0141386698656debe691d33b9e1f90570f42b0cd7a244cd9fde"} Feb 18 14:58:56 crc kubenswrapper[4896]: I0218 14:58:56.125108 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wzwx5" Feb 18 14:58:56 crc kubenswrapper[4896]: I0218 14:58:56.139598 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wzwx5"] Feb 18 14:58:56 crc kubenswrapper[4896]: I0218 14:58:56.142095 4896 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wzwx5"] Feb 18 14:58:57 crc kubenswrapper[4896]: I0218 14:58:57.934737 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f48b5c5e-4ec2-4bca-bd26-02b837c029ea" path="/var/lib/kubelet/pods/f48b5c5e-4ec2-4bca-bd26-02b837c029ea/volumes" Feb 18 14:58:58 crc kubenswrapper[4896]: I0218 14:58:58.820951 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t6sfb"] Feb 18 14:58:58 crc kubenswrapper[4896]: I0218 14:58:58.821498 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-t6sfb" podUID="eddbcdff-a387-4f95-9277-5850b6485036" containerName="registry-server" containerID="cri-o://6a9b0468659ba66ab625173e6087c0b54094b68f2dc654936d5502396abaa27c" gracePeriod=2 Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.149190 4896 generic.go:334] "Generic (PLEG): container finished" podID="eddbcdff-a387-4f95-9277-5850b6485036" containerID="6a9b0468659ba66ab625173e6087c0b54094b68f2dc654936d5502396abaa27c" exitCode=0 Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.149255 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t6sfb" event={"ID":"eddbcdff-a387-4f95-9277-5850b6485036","Type":"ContainerDied","Data":"6a9b0468659ba66ab625173e6087c0b54094b68f2dc654936d5502396abaa27c"} Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.169471 4896 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 18 14:58:59 crc kubenswrapper[4896]: E0218 14:58:59.169866 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfcee59f-69ac-42ac-86e8-20c08066d0c1" containerName="pruner" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.169877 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfcee59f-69ac-42ac-86e8-20c08066d0c1" containerName="pruner" Feb 18 14:58:59 crc kubenswrapper[4896]: E0218 14:58:59.169889 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f48b5c5e-4ec2-4bca-bd26-02b837c029ea" containerName="extract-utilities" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.169897 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="f48b5c5e-4ec2-4bca-bd26-02b837c029ea" containerName="extract-utilities" Feb 18 14:58:59 crc kubenswrapper[4896]: E0218 14:58:59.169905 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f48b5c5e-4ec2-4bca-bd26-02b837c029ea" containerName="registry-server" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.169911 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="f48b5c5e-4ec2-4bca-bd26-02b837c029ea" containerName="registry-server" Feb 18 14:58:59 crc kubenswrapper[4896]: E0218 14:58:59.169920 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f48b5c5e-4ec2-4bca-bd26-02b837c029ea" containerName="extract-content" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.169926 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="f48b5c5e-4ec2-4bca-bd26-02b837c029ea" containerName="extract-content" Feb 18 14:58:59 crc kubenswrapper[4896]: E0218 14:58:59.169933 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c93c74b9-3b90-4f6d-ab3b-8210d0704ce0" containerName="registry-server" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.169938 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="c93c74b9-3b90-4f6d-ab3b-8210d0704ce0" containerName="registry-server" Feb 18 14:58:59 crc kubenswrapper[4896]: E0218 14:58:59.169950 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c93c74b9-3b90-4f6d-ab3b-8210d0704ce0" containerName="extract-content" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.169955 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="c93c74b9-3b90-4f6d-ab3b-8210d0704ce0" containerName="extract-content" Feb 18 14:58:59 crc kubenswrapper[4896]: E0218 14:58:59.169963 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c93c74b9-3b90-4f6d-ab3b-8210d0704ce0" containerName="extract-utilities" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.169968 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="c93c74b9-3b90-4f6d-ab3b-8210d0704ce0" containerName="extract-utilities" Feb 18 14:58:59 crc kubenswrapper[4896]: E0218 14:58:59.169974 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdc2c09e-a873-4f65-8a5f-949b4562cb10" containerName="registry-server" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.169980 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdc2c09e-a873-4f65-8a5f-949b4562cb10" containerName="registry-server" Feb 18 14:58:59 crc kubenswrapper[4896]: E0218 14:58:59.169987 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdc2c09e-a873-4f65-8a5f-949b4562cb10" containerName="extract-utilities" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.169992 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdc2c09e-a873-4f65-8a5f-949b4562cb10" containerName="extract-utilities" Feb 18 14:58:59 crc kubenswrapper[4896]: E0218 14:58:59.170001 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdc2c09e-a873-4f65-8a5f-949b4562cb10" containerName="extract-content" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.170006 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdc2c09e-a873-4f65-8a5f-949b4562cb10" containerName="extract-content" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.170132 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfcee59f-69ac-42ac-86e8-20c08066d0c1" containerName="pruner" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.170142 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdc2c09e-a873-4f65-8a5f-949b4562cb10" containerName="registry-server" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.170154 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="f48b5c5e-4ec2-4bca-bd26-02b837c029ea" containerName="registry-server" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.170163 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="c93c74b9-3b90-4f6d-ab3b-8210d0704ce0" containerName="registry-server" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.170494 4896 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.170735 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed" gracePeriod=15 Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.170866 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.171187 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://646e1cf96a28a07565ec2d56a4ffc32c10e95356b7ebf2399b6b3ba833274a68" gracePeriod=15 Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.171284 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d" gracePeriod=15 Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.171345 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f" gracePeriod=15 Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.171342 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21" gracePeriod=15 Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.172256 4896 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 18 14:58:59 crc kubenswrapper[4896]: E0218 14:58:59.172374 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.172384 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 18 14:58:59 crc kubenswrapper[4896]: E0218 14:58:59.172392 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.172398 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 18 14:58:59 crc kubenswrapper[4896]: E0218 14:58:59.172405 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.172410 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 18 14:58:59 crc kubenswrapper[4896]: E0218 14:58:59.172423 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.172429 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 18 14:58:59 crc kubenswrapper[4896]: E0218 14:58:59.172439 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.172446 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 18 14:58:59 crc kubenswrapper[4896]: E0218 14:58:59.172453 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.172459 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 18 14:58:59 crc kubenswrapper[4896]: E0218 14:58:59.172464 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.172470 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 18 14:58:59 crc kubenswrapper[4896]: E0218 14:58:59.172477 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.172489 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.172571 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.172580 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.172592 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.172600 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.172638 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.172647 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.172809 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 18 14:58:59 crc kubenswrapper[4896]: E0218 14:58:59.252093 4896 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.156:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.255924 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t6sfb" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.256461 4896 status_manager.go:851] "Failed to get status for pod" podUID="eddbcdff-a387-4f95-9277-5850b6485036" pod="openshift-marketplace/redhat-operators-t6sfb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-t6sfb\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.256731 4896 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.295153 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.295248 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.295336 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.295378 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.295508 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.295575 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.295606 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.295627 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.395941 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eddbcdff-a387-4f95-9277-5850b6485036-catalog-content\") pod \"eddbcdff-a387-4f95-9277-5850b6485036\" (UID: \"eddbcdff-a387-4f95-9277-5850b6485036\") " Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.396298 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dc4l9\" (UniqueName: \"kubernetes.io/projected/eddbcdff-a387-4f95-9277-5850b6485036-kube-api-access-dc4l9\") pod \"eddbcdff-a387-4f95-9277-5850b6485036\" (UID: \"eddbcdff-a387-4f95-9277-5850b6485036\") " Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.396365 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eddbcdff-a387-4f95-9277-5850b6485036-utilities\") pod \"eddbcdff-a387-4f95-9277-5850b6485036\" (UID: \"eddbcdff-a387-4f95-9277-5850b6485036\") " Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.396677 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.396759 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.397210 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eddbcdff-a387-4f95-9277-5850b6485036-utilities" (OuterVolumeSpecName: "utilities") pod "eddbcdff-a387-4f95-9277-5850b6485036" (UID: "eddbcdff-a387-4f95-9277-5850b6485036"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.397666 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.397772 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.397800 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.397848 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.397864 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.397903 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.397927 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.397982 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.398011 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.398051 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.398139 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.398154 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.398182 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.398269 4896 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eddbcdff-a387-4f95-9277-5850b6485036-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.398318 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.401457 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eddbcdff-a387-4f95-9277-5850b6485036-kube-api-access-dc4l9" (OuterVolumeSpecName: "kube-api-access-dc4l9") pod "eddbcdff-a387-4f95-9277-5850b6485036" (UID: "eddbcdff-a387-4f95-9277-5850b6485036"). InnerVolumeSpecName "kube-api-access-dc4l9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.499748 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dc4l9\" (UniqueName: \"kubernetes.io/projected/eddbcdff-a387-4f95-9277-5850b6485036-kube-api-access-dc4l9\") on node \"crc\" DevicePath \"\"" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.509037 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eddbcdff-a387-4f95-9277-5850b6485036-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eddbcdff-a387-4f95-9277-5850b6485036" (UID: "eddbcdff-a387-4f95-9277-5850b6485036"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.553639 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 14:58:59 crc kubenswrapper[4896]: W0218 14:58:59.582388 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-6a94c92fabba5a39a9e424ba24b0cc9f08bb3b5c6974f978a7894ab5a1c7b368 WatchSource:0}: Error finding container 6a94c92fabba5a39a9e424ba24b0cc9f08bb3b5c6974f978a7894ab5a1c7b368: Status 404 returned error can't find the container with id 6a94c92fabba5a39a9e424ba24b0cc9f08bb3b5c6974f978a7894ab5a1c7b368 Feb 18 14:58:59 crc kubenswrapper[4896]: E0218 14:58:59.585672 4896 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.156:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.18955f3bf697d0fe openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-18 14:58:59.585011966 +0000 UTC m=+238.514346179,LastTimestamp:2026-02-18 14:58:59.585011966 +0000 UTC m=+238.514346179,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 18 14:58:59 crc kubenswrapper[4896]: I0218 14:58:59.600276 4896 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eddbcdff-a387-4f95-9277-5850b6485036-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:00 crc kubenswrapper[4896]: I0218 14:59:00.155518 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t6sfb" event={"ID":"eddbcdff-a387-4f95-9277-5850b6485036","Type":"ContainerDied","Data":"ffde705eb9656db2f46d1850ecd5a7172d23eb4b4a8313ffe1f197d209e52cf4"} Feb 18 14:59:00 crc kubenswrapper[4896]: I0218 14:59:00.155566 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t6sfb" Feb 18 14:59:00 crc kubenswrapper[4896]: I0218 14:59:00.155584 4896 scope.go:117] "RemoveContainer" containerID="6a9b0468659ba66ab625173e6087c0b54094b68f2dc654936d5502396abaa27c" Feb 18 14:59:00 crc kubenswrapper[4896]: I0218 14:59:00.156315 4896 status_manager.go:851] "Failed to get status for pod" podUID="eddbcdff-a387-4f95-9277-5850b6485036" pod="openshift-marketplace/redhat-operators-t6sfb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-t6sfb\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:00 crc kubenswrapper[4896]: I0218 14:59:00.156960 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"df25631ac2f71f0842be433055c01f20cb4d72be8041d8d32c5ff0b4cdd6866c"} Feb 18 14:59:00 crc kubenswrapper[4896]: I0218 14:59:00.156985 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"6a94c92fabba5a39a9e424ba24b0cc9f08bb3b5c6974f978a7894ab5a1c7b368"} Feb 18 14:59:00 crc kubenswrapper[4896]: E0218 14:59:00.157464 4896 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.156:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 14:59:00 crc kubenswrapper[4896]: I0218 14:59:00.157521 4896 status_manager.go:851] "Failed to get status for pod" podUID="eddbcdff-a387-4f95-9277-5850b6485036" pod="openshift-marketplace/redhat-operators-t6sfb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-t6sfb\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:00 crc kubenswrapper[4896]: I0218 14:59:00.159048 4896 generic.go:334] "Generic (PLEG): container finished" podID="4a847ad6-23c1-43e2-b916-c97df45c2aff" containerID="00702078018dc14dc016ecccfdb84e35551d242eec8598812df3356279f9e740" exitCode=0 Feb 18 14:59:00 crc kubenswrapper[4896]: I0218 14:59:00.159106 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"4a847ad6-23c1-43e2-b916-c97df45c2aff","Type":"ContainerDied","Data":"00702078018dc14dc016ecccfdb84e35551d242eec8598812df3356279f9e740"} Feb 18 14:59:00 crc kubenswrapper[4896]: I0218 14:59:00.159501 4896 status_manager.go:851] "Failed to get status for pod" podUID="4a847ad6-23c1-43e2-b916-c97df45c2aff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:00 crc kubenswrapper[4896]: I0218 14:59:00.159951 4896 status_manager.go:851] "Failed to get status for pod" podUID="eddbcdff-a387-4f95-9277-5850b6485036" pod="openshift-marketplace/redhat-operators-t6sfb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-t6sfb\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:00 crc kubenswrapper[4896]: I0218 14:59:00.160571 4896 status_manager.go:851] "Failed to get status for pod" podUID="4a847ad6-23c1-43e2-b916-c97df45c2aff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:00 crc kubenswrapper[4896]: I0218 14:59:00.160888 4896 status_manager.go:851] "Failed to get status for pod" podUID="eddbcdff-a387-4f95-9277-5850b6485036" pod="openshift-marketplace/redhat-operators-t6sfb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-t6sfb\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:00 crc kubenswrapper[4896]: I0218 14:59:00.161059 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 18 14:59:00 crc kubenswrapper[4896]: I0218 14:59:00.162745 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 18 14:59:00 crc kubenswrapper[4896]: I0218 14:59:00.163377 4896 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="646e1cf96a28a07565ec2d56a4ffc32c10e95356b7ebf2399b6b3ba833274a68" exitCode=0 Feb 18 14:59:00 crc kubenswrapper[4896]: I0218 14:59:00.163401 4896 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f" exitCode=0 Feb 18 14:59:00 crc kubenswrapper[4896]: I0218 14:59:00.163412 4896 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d" exitCode=0 Feb 18 14:59:00 crc kubenswrapper[4896]: I0218 14:59:00.163420 4896 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21" exitCode=2 Feb 18 14:59:00 crc kubenswrapper[4896]: I0218 14:59:00.168291 4896 scope.go:117] "RemoveContainer" containerID="d943c2184f788626ab04198b6da0477b59006414dee16d46906e39d9fb3600f0" Feb 18 14:59:00 crc kubenswrapper[4896]: I0218 14:59:00.184148 4896 scope.go:117] "RemoveContainer" containerID="9bc72d58d15b9d44e16885e79087167612fa6753010eba7d950dd37efe2cf549" Feb 18 14:59:00 crc kubenswrapper[4896]: I0218 14:59:00.201203 4896 scope.go:117] "RemoveContainer" containerID="fab18f6a5c83d4b1ae692eac24d97103d6fd6f96ada424a7231770fd9babad23" Feb 18 14:59:01 crc kubenswrapper[4896]: E0218 14:59:01.015168 4896 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.156:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" volumeName="registry-storage" Feb 18 14:59:01 crc kubenswrapper[4896]: I0218 14:59:01.173188 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 18 14:59:01 crc kubenswrapper[4896]: I0218 14:59:01.564486 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 18 14:59:01 crc kubenswrapper[4896]: I0218 14:59:01.564937 4896 status_manager.go:851] "Failed to get status for pod" podUID="eddbcdff-a387-4f95-9277-5850b6485036" pod="openshift-marketplace/redhat-operators-t6sfb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-t6sfb\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:01 crc kubenswrapper[4896]: I0218 14:59:01.565117 4896 status_manager.go:851] "Failed to get status for pod" podUID="4a847ad6-23c1-43e2-b916-c97df45c2aff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:01 crc kubenswrapper[4896]: I0218 14:59:01.619643 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 18 14:59:01 crc kubenswrapper[4896]: I0218 14:59:01.620431 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:59:01 crc kubenswrapper[4896]: I0218 14:59:01.621522 4896 status_manager.go:851] "Failed to get status for pod" podUID="eddbcdff-a387-4f95-9277-5850b6485036" pod="openshift-marketplace/redhat-operators-t6sfb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-t6sfb\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:01 crc kubenswrapper[4896]: I0218 14:59:01.621740 4896 status_manager.go:851] "Failed to get status for pod" podUID="4a847ad6-23c1-43e2-b916-c97df45c2aff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:01 crc kubenswrapper[4896]: I0218 14:59:01.622009 4896 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:01 crc kubenswrapper[4896]: I0218 14:59:01.626281 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 18 14:59:01 crc kubenswrapper[4896]: I0218 14:59:01.626349 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4a847ad6-23c1-43e2-b916-c97df45c2aff-kubelet-dir\") pod \"4a847ad6-23c1-43e2-b916-c97df45c2aff\" (UID: \"4a847ad6-23c1-43e2-b916-c97df45c2aff\") " Feb 18 14:59:01 crc kubenswrapper[4896]: I0218 14:59:01.626375 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 18 14:59:01 crc kubenswrapper[4896]: I0218 14:59:01.626394 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 18 14:59:01 crc kubenswrapper[4896]: I0218 14:59:01.626409 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/4a847ad6-23c1-43e2-b916-c97df45c2aff-var-lock\") pod \"4a847ad6-23c1-43e2-b916-c97df45c2aff\" (UID: \"4a847ad6-23c1-43e2-b916-c97df45c2aff\") " Feb 18 14:59:01 crc kubenswrapper[4896]: I0218 14:59:01.626438 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4a847ad6-23c1-43e2-b916-c97df45c2aff-kube-api-access\") pod \"4a847ad6-23c1-43e2-b916-c97df45c2aff\" (UID: \"4a847ad6-23c1-43e2-b916-c97df45c2aff\") " Feb 18 14:59:01 crc kubenswrapper[4896]: I0218 14:59:01.627262 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 14:59:01 crc kubenswrapper[4896]: I0218 14:59:01.627300 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4a847ad6-23c1-43e2-b916-c97df45c2aff-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "4a847ad6-23c1-43e2-b916-c97df45c2aff" (UID: "4a847ad6-23c1-43e2-b916-c97df45c2aff"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 14:59:01 crc kubenswrapper[4896]: I0218 14:59:01.627320 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 14:59:01 crc kubenswrapper[4896]: I0218 14:59:01.627336 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 14:59:01 crc kubenswrapper[4896]: I0218 14:59:01.627352 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4a847ad6-23c1-43e2-b916-c97df45c2aff-var-lock" (OuterVolumeSpecName: "var-lock") pod "4a847ad6-23c1-43e2-b916-c97df45c2aff" (UID: "4a847ad6-23c1-43e2-b916-c97df45c2aff"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 14:59:01 crc kubenswrapper[4896]: I0218 14:59:01.633435 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a847ad6-23c1-43e2-b916-c97df45c2aff-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "4a847ad6-23c1-43e2-b916-c97df45c2aff" (UID: "4a847ad6-23c1-43e2-b916-c97df45c2aff"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:59:01 crc kubenswrapper[4896]: I0218 14:59:01.727364 4896 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4a847ad6-23c1-43e2-b916-c97df45c2aff-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:01 crc kubenswrapper[4896]: I0218 14:59:01.727390 4896 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:01 crc kubenswrapper[4896]: I0218 14:59:01.727400 4896 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:01 crc kubenswrapper[4896]: I0218 14:59:01.727412 4896 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/4a847ad6-23c1-43e2-b916-c97df45c2aff-var-lock\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:01 crc kubenswrapper[4896]: I0218 14:59:01.727421 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4a847ad6-23c1-43e2-b916-c97df45c2aff-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:01 crc kubenswrapper[4896]: I0218 14:59:01.727430 4896 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:01 crc kubenswrapper[4896]: I0218 14:59:01.930542 4896 status_manager.go:851] "Failed to get status for pod" podUID="eddbcdff-a387-4f95-9277-5850b6485036" pod="openshift-marketplace/redhat-operators-t6sfb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-t6sfb\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:01 crc kubenswrapper[4896]: I0218 14:59:01.930922 4896 status_manager.go:851] "Failed to get status for pod" podUID="4a847ad6-23c1-43e2-b916-c97df45c2aff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:01 crc kubenswrapper[4896]: I0218 14:59:01.931213 4896 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:01 crc kubenswrapper[4896]: I0218 14:59:01.934144 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Feb 18 14:59:02 crc kubenswrapper[4896]: I0218 14:59:02.180589 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 18 14:59:02 crc kubenswrapper[4896]: I0218 14:59:02.180742 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"4a847ad6-23c1-43e2-b916-c97df45c2aff","Type":"ContainerDied","Data":"bf35330fd57e30497c9c43ee40c3f185dc86c6a6034b62d94debd3f52ec6439f"} Feb 18 14:59:02 crc kubenswrapper[4896]: I0218 14:59:02.181231 4896 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf35330fd57e30497c9c43ee40c3f185dc86c6a6034b62d94debd3f52ec6439f" Feb 18 14:59:02 crc kubenswrapper[4896]: I0218 14:59:02.184433 4896 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed" exitCode=0 Feb 18 14:59:02 crc kubenswrapper[4896]: I0218 14:59:02.184493 4896 scope.go:117] "RemoveContainer" containerID="646e1cf96a28a07565ec2d56a4ffc32c10e95356b7ebf2399b6b3ba833274a68" Feb 18 14:59:02 crc kubenswrapper[4896]: I0218 14:59:02.184495 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:59:02 crc kubenswrapper[4896]: I0218 14:59:02.184529 4896 status_manager.go:851] "Failed to get status for pod" podUID="4a847ad6-23c1-43e2-b916-c97df45c2aff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:02 crc kubenswrapper[4896]: I0218 14:59:02.184925 4896 status_manager.go:851] "Failed to get status for pod" podUID="eddbcdff-a387-4f95-9277-5850b6485036" pod="openshift-marketplace/redhat-operators-t6sfb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-t6sfb\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:02 crc kubenswrapper[4896]: I0218 14:59:02.185184 4896 status_manager.go:851] "Failed to get status for pod" podUID="4a847ad6-23c1-43e2-b916-c97df45c2aff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:02 crc kubenswrapper[4896]: I0218 14:59:02.185445 4896 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:02 crc kubenswrapper[4896]: I0218 14:59:02.185699 4896 status_manager.go:851] "Failed to get status for pod" podUID="eddbcdff-a387-4f95-9277-5850b6485036" pod="openshift-marketplace/redhat-operators-t6sfb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-t6sfb\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:02 crc kubenswrapper[4896]: I0218 14:59:02.186437 4896 status_manager.go:851] "Failed to get status for pod" podUID="eddbcdff-a387-4f95-9277-5850b6485036" pod="openshift-marketplace/redhat-operators-t6sfb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-t6sfb\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:02 crc kubenswrapper[4896]: I0218 14:59:02.186729 4896 status_manager.go:851] "Failed to get status for pod" podUID="4a847ad6-23c1-43e2-b916-c97df45c2aff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:02 crc kubenswrapper[4896]: I0218 14:59:02.186923 4896 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:02 crc kubenswrapper[4896]: I0218 14:59:02.199637 4896 scope.go:117] "RemoveContainer" containerID="207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f" Feb 18 14:59:02 crc kubenswrapper[4896]: I0218 14:59:02.212054 4896 scope.go:117] "RemoveContainer" containerID="abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d" Feb 18 14:59:02 crc kubenswrapper[4896]: I0218 14:59:02.225940 4896 scope.go:117] "RemoveContainer" containerID="1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21" Feb 18 14:59:02 crc kubenswrapper[4896]: I0218 14:59:02.237666 4896 scope.go:117] "RemoveContainer" containerID="42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed" Feb 18 14:59:02 crc kubenswrapper[4896]: I0218 14:59:02.257600 4896 scope.go:117] "RemoveContainer" containerID="f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38" Feb 18 14:59:02 crc kubenswrapper[4896]: I0218 14:59:02.272315 4896 scope.go:117] "RemoveContainer" containerID="646e1cf96a28a07565ec2d56a4ffc32c10e95356b7ebf2399b6b3ba833274a68" Feb 18 14:59:02 crc kubenswrapper[4896]: E0218 14:59:02.272661 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"646e1cf96a28a07565ec2d56a4ffc32c10e95356b7ebf2399b6b3ba833274a68\": container with ID starting with 646e1cf96a28a07565ec2d56a4ffc32c10e95356b7ebf2399b6b3ba833274a68 not found: ID does not exist" containerID="646e1cf96a28a07565ec2d56a4ffc32c10e95356b7ebf2399b6b3ba833274a68" Feb 18 14:59:02 crc kubenswrapper[4896]: I0218 14:59:02.272696 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"646e1cf96a28a07565ec2d56a4ffc32c10e95356b7ebf2399b6b3ba833274a68"} err="failed to get container status \"646e1cf96a28a07565ec2d56a4ffc32c10e95356b7ebf2399b6b3ba833274a68\": rpc error: code = NotFound desc = could not find container \"646e1cf96a28a07565ec2d56a4ffc32c10e95356b7ebf2399b6b3ba833274a68\": container with ID starting with 646e1cf96a28a07565ec2d56a4ffc32c10e95356b7ebf2399b6b3ba833274a68 not found: ID does not exist" Feb 18 14:59:02 crc kubenswrapper[4896]: I0218 14:59:02.272723 4896 scope.go:117] "RemoveContainer" containerID="207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f" Feb 18 14:59:02 crc kubenswrapper[4896]: E0218 14:59:02.273106 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\": container with ID starting with 207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f not found: ID does not exist" containerID="207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f" Feb 18 14:59:02 crc kubenswrapper[4896]: I0218 14:59:02.273133 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f"} err="failed to get container status \"207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\": rpc error: code = NotFound desc = could not find container \"207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f\": container with ID starting with 207ab2a9cf480a0d7e28b63d9c91e62d5bd20c80ed2c098dd4dc58aacaecd52f not found: ID does not exist" Feb 18 14:59:02 crc kubenswrapper[4896]: I0218 14:59:02.273154 4896 scope.go:117] "RemoveContainer" containerID="abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d" Feb 18 14:59:02 crc kubenswrapper[4896]: E0218 14:59:02.273709 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\": container with ID starting with abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d not found: ID does not exist" containerID="abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d" Feb 18 14:59:02 crc kubenswrapper[4896]: I0218 14:59:02.273729 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d"} err="failed to get container status \"abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\": rpc error: code = NotFound desc = could not find container \"abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d\": container with ID starting with abfcd271344790f9fff09d606943ac2dde86f48abc9cbf58eb423fd18a81fb7d not found: ID does not exist" Feb 18 14:59:02 crc kubenswrapper[4896]: I0218 14:59:02.273743 4896 scope.go:117] "RemoveContainer" containerID="1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21" Feb 18 14:59:02 crc kubenswrapper[4896]: E0218 14:59:02.274071 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\": container with ID starting with 1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21 not found: ID does not exist" containerID="1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21" Feb 18 14:59:02 crc kubenswrapper[4896]: I0218 14:59:02.274123 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21"} err="failed to get container status \"1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\": rpc error: code = NotFound desc = could not find container \"1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21\": container with ID starting with 1b5d7b8bc143beac0258690bcc62b8d17890fcdd4417db01890bc942da52ae21 not found: ID does not exist" Feb 18 14:59:02 crc kubenswrapper[4896]: I0218 14:59:02.274159 4896 scope.go:117] "RemoveContainer" containerID="42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed" Feb 18 14:59:02 crc kubenswrapper[4896]: E0218 14:59:02.274462 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\": container with ID starting with 42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed not found: ID does not exist" containerID="42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed" Feb 18 14:59:02 crc kubenswrapper[4896]: I0218 14:59:02.274481 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed"} err="failed to get container status \"42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\": rpc error: code = NotFound desc = could not find container \"42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed\": container with ID starting with 42940b4e9660f1b467ac20184408379e22e39d1c5b091f3021f2fb247e45efed not found: ID does not exist" Feb 18 14:59:02 crc kubenswrapper[4896]: I0218 14:59:02.274494 4896 scope.go:117] "RemoveContainer" containerID="f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38" Feb 18 14:59:02 crc kubenswrapper[4896]: E0218 14:59:02.274758 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\": container with ID starting with f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38 not found: ID does not exist" containerID="f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38" Feb 18 14:59:02 crc kubenswrapper[4896]: I0218 14:59:02.274780 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38"} err="failed to get container status \"f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\": rpc error: code = NotFound desc = could not find container \"f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38\": container with ID starting with f93c7291e3992490a8a7dcd370a316c252acf0b4936f73d6b685f5c6b8054b38 not found: ID does not exist" Feb 18 14:59:03 crc kubenswrapper[4896]: E0218 14:59:03.257567 4896 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:03 crc kubenswrapper[4896]: E0218 14:59:03.258019 4896 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:03 crc kubenswrapper[4896]: E0218 14:59:03.258259 4896 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:03 crc kubenswrapper[4896]: E0218 14:59:03.258429 4896 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:03 crc kubenswrapper[4896]: E0218 14:59:03.258614 4896 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:03 crc kubenswrapper[4896]: I0218 14:59:03.258642 4896 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Feb 18 14:59:03 crc kubenswrapper[4896]: E0218 14:59:03.258857 4896 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.156:6443: connect: connection refused" interval="200ms" Feb 18 14:59:03 crc kubenswrapper[4896]: E0218 14:59:03.459516 4896 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.156:6443: connect: connection refused" interval="400ms" Feb 18 14:59:03 crc kubenswrapper[4896]: E0218 14:59:03.860718 4896 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.156:6443: connect: connection refused" interval="800ms" Feb 18 14:59:04 crc kubenswrapper[4896]: E0218 14:59:04.287813 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:59:04Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:59:04Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:59:04Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-18T14:59:04Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:04 crc kubenswrapper[4896]: E0218 14:59:04.288514 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:04 crc kubenswrapper[4896]: E0218 14:59:04.288969 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:04 crc kubenswrapper[4896]: E0218 14:59:04.289337 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:04 crc kubenswrapper[4896]: E0218 14:59:04.289556 4896 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:04 crc kubenswrapper[4896]: E0218 14:59:04.289579 4896 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 18 14:59:04 crc kubenswrapper[4896]: E0218 14:59:04.661375 4896 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.156:6443: connect: connection refused" interval="1.6s" Feb 18 14:59:06 crc kubenswrapper[4896]: E0218 14:59:06.262839 4896 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.156:6443: connect: connection refused" interval="3.2s" Feb 18 14:59:09 crc kubenswrapper[4896]: E0218 14:59:09.036491 4896 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.156:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.18955f3bf697d0fe openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-18 14:58:59.585011966 +0000 UTC m=+238.514346179,LastTimestamp:2026-02-18 14:58:59.585011966 +0000 UTC m=+238.514346179,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 18 14:59:09 crc kubenswrapper[4896]: E0218 14:59:09.463685 4896 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.156:6443: connect: connection refused" interval="6.4s" Feb 18 14:59:10 crc kubenswrapper[4896]: I0218 14:59:10.927025 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:59:10 crc kubenswrapper[4896]: I0218 14:59:10.927675 4896 status_manager.go:851] "Failed to get status for pod" podUID="eddbcdff-a387-4f95-9277-5850b6485036" pod="openshift-marketplace/redhat-operators-t6sfb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-t6sfb\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:10 crc kubenswrapper[4896]: I0218 14:59:10.928010 4896 status_manager.go:851] "Failed to get status for pod" podUID="4a847ad6-23c1-43e2-b916-c97df45c2aff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:10 crc kubenswrapper[4896]: I0218 14:59:10.948810 4896 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e1e5f7db-b3a9-44d9-8ab1-1950995db98a" Feb 18 14:59:10 crc kubenswrapper[4896]: I0218 14:59:10.948842 4896 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e1e5f7db-b3a9-44d9-8ab1-1950995db98a" Feb 18 14:59:10 crc kubenswrapper[4896]: E0218 14:59:10.949192 4896 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.156:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:59:10 crc kubenswrapper[4896]: I0218 14:59:10.949776 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:59:11 crc kubenswrapper[4896]: I0218 14:59:11.229842 4896 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="d30b097dfbac66414895545659c7c64a81821bb3e314ab6914ef637b221407c8" exitCode=0 Feb 18 14:59:11 crc kubenswrapper[4896]: I0218 14:59:11.229934 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"d30b097dfbac66414895545659c7c64a81821bb3e314ab6914ef637b221407c8"} Feb 18 14:59:11 crc kubenswrapper[4896]: I0218 14:59:11.230194 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"8fe15b05f96c7ed3b6b32aef6abcb01742f3fb1dba297e0cdbb5ac4c1923da14"} Feb 18 14:59:11 crc kubenswrapper[4896]: I0218 14:59:11.230659 4896 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e1e5f7db-b3a9-44d9-8ab1-1950995db98a" Feb 18 14:59:11 crc kubenswrapper[4896]: I0218 14:59:11.230680 4896 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e1e5f7db-b3a9-44d9-8ab1-1950995db98a" Feb 18 14:59:11 crc kubenswrapper[4896]: E0218 14:59:11.231106 4896 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.156:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:59:11 crc kubenswrapper[4896]: I0218 14:59:11.231137 4896 status_manager.go:851] "Failed to get status for pod" podUID="eddbcdff-a387-4f95-9277-5850b6485036" pod="openshift-marketplace/redhat-operators-t6sfb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-t6sfb\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:11 crc kubenswrapper[4896]: I0218 14:59:11.231468 4896 status_manager.go:851] "Failed to get status for pod" podUID="4a847ad6-23c1-43e2-b916-c97df45c2aff" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.156:6443: connect: connection refused" Feb 18 14:59:12 crc kubenswrapper[4896]: I0218 14:59:12.238720 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 18 14:59:12 crc kubenswrapper[4896]: I0218 14:59:12.238962 4896 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8" exitCode=1 Feb 18 14:59:12 crc kubenswrapper[4896]: I0218 14:59:12.239007 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8"} Feb 18 14:59:12 crc kubenswrapper[4896]: I0218 14:59:12.239409 4896 scope.go:117] "RemoveContainer" containerID="dbf6d7601ff4717fc73e468b9998a825ddde900ccc0319e2e94b53b8c055ffc8" Feb 18 14:59:12 crc kubenswrapper[4896]: I0218 14:59:12.243060 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"de01f4b1327ea5ba9b1e7ca63be1de3be858cd624c5b7904d5375c0438e8c4f7"} Feb 18 14:59:12 crc kubenswrapper[4896]: I0218 14:59:12.243089 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"fc047bd6cb211834b4d58d066464ad85cce46d85fa1a529315711378567e321a"} Feb 18 14:59:12 crc kubenswrapper[4896]: I0218 14:59:12.243098 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"877f349c74ad7912dd8c3f87ab19f0681e5d94ca8ffc39adb7435200613dba85"} Feb 18 14:59:12 crc kubenswrapper[4896]: I0218 14:59:12.243106 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"106cc83e0b0e52a7e326e5ee88522bff210e6e1ea4cbbd6d27d250dab4b8b3eb"} Feb 18 14:59:12 crc kubenswrapper[4896]: I0218 14:59:12.243115 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"7475cde6173fc1d3d5001e9c0b6e88a5614630b681e02671e592d9e339e21d0d"} Feb 18 14:59:12 crc kubenswrapper[4896]: I0218 14:59:12.243352 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:59:12 crc kubenswrapper[4896]: I0218 14:59:12.243396 4896 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e1e5f7db-b3a9-44d9-8ab1-1950995db98a" Feb 18 14:59:12 crc kubenswrapper[4896]: I0218 14:59:12.243413 4896 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e1e5f7db-b3a9-44d9-8ab1-1950995db98a" Feb 18 14:59:13 crc kubenswrapper[4896]: I0218 14:59:13.289023 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 18 14:59:13 crc kubenswrapper[4896]: I0218 14:59:13.289083 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7215be5a359c3ceea59fb61360be134fa952f2a5027c704695f2b38dfadbfc55"} Feb 18 14:59:14 crc kubenswrapper[4896]: I0218 14:59:14.955001 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 14:59:14 crc kubenswrapper[4896]: I0218 14:59:14.958596 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 14:59:15 crc kubenswrapper[4896]: I0218 14:59:15.298519 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 14:59:15 crc kubenswrapper[4896]: I0218 14:59:15.950879 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:59:15 crc kubenswrapper[4896]: I0218 14:59:15.951204 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:59:15 crc kubenswrapper[4896]: I0218 14:59:15.955600 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:59:17 crc kubenswrapper[4896]: I0218 14:59:17.751998 4896 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:59:17 crc kubenswrapper[4896]: I0218 14:59:17.842925 4896 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="11d2c1a6-b51b-4681-a124-78a15d07c266" Feb 18 14:59:18 crc kubenswrapper[4896]: I0218 14:59:18.314184 4896 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e1e5f7db-b3a9-44d9-8ab1-1950995db98a" Feb 18 14:59:18 crc kubenswrapper[4896]: I0218 14:59:18.314490 4896 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="e1e5f7db-b3a9-44d9-8ab1-1950995db98a" Feb 18 14:59:18 crc kubenswrapper[4896]: I0218 14:59:18.316046 4896 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="11d2c1a6-b51b-4681-a124-78a15d07c266" Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.214899 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" podUID="76cb1e0e-4321-4450-9fcb-b505a7a1d969" containerName="oauth-openshift" containerID="cri-o://fce8bf5bd3d671ffcdafbfabccda9c929d8b1b87b96a506cadf5a5c962411ee1" gracePeriod=15 Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.768599 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.844189 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-service-ca\") pod \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.844323 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-user-idp-0-file-data\") pod \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.844364 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-router-certs\") pod \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.844392 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-user-template-provider-selection\") pod \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.844425 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-cliconfig\") pod \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.844456 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-session\") pod \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.844481 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6pk7\" (UniqueName: \"kubernetes.io/projected/76cb1e0e-4321-4450-9fcb-b505a7a1d969-kube-api-access-q6pk7\") pod \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.844517 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-ocp-branding-template\") pod \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.844547 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-user-template-error\") pod \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.845161 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "76cb1e0e-4321-4450-9fcb-b505a7a1d969" (UID: "76cb1e0e-4321-4450-9fcb-b505a7a1d969"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.845338 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/76cb1e0e-4321-4450-9fcb-b505a7a1d969-audit-policies\") pod \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.845376 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-serving-cert\") pod \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.845412 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-user-template-login\") pod \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.845434 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/76cb1e0e-4321-4450-9fcb-b505a7a1d969-audit-dir\") pod \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.845452 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-trusted-ca-bundle\") pod \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\" (UID: \"76cb1e0e-4321-4450-9fcb-b505a7a1d969\") " Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.846010 4896 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.846479 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/76cb1e0e-4321-4450-9fcb-b505a7a1d969-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "76cb1e0e-4321-4450-9fcb-b505a7a1d969" (UID: "76cb1e0e-4321-4450-9fcb-b505a7a1d969"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.847879 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "76cb1e0e-4321-4450-9fcb-b505a7a1d969" (UID: "76cb1e0e-4321-4450-9fcb-b505a7a1d969"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.848531 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76cb1e0e-4321-4450-9fcb-b505a7a1d969-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "76cb1e0e-4321-4450-9fcb-b505a7a1d969" (UID: "76cb1e0e-4321-4450-9fcb-b505a7a1d969"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.848932 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "76cb1e0e-4321-4450-9fcb-b505a7a1d969" (UID: "76cb1e0e-4321-4450-9fcb-b505a7a1d969"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.851037 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "76cb1e0e-4321-4450-9fcb-b505a7a1d969" (UID: "76cb1e0e-4321-4450-9fcb-b505a7a1d969"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.852002 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "76cb1e0e-4321-4450-9fcb-b505a7a1d969" (UID: "76cb1e0e-4321-4450-9fcb-b505a7a1d969"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.856011 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "76cb1e0e-4321-4450-9fcb-b505a7a1d969" (UID: "76cb1e0e-4321-4450-9fcb-b505a7a1d969"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.856026 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76cb1e0e-4321-4450-9fcb-b505a7a1d969-kube-api-access-q6pk7" (OuterVolumeSpecName: "kube-api-access-q6pk7") pod "76cb1e0e-4321-4450-9fcb-b505a7a1d969" (UID: "76cb1e0e-4321-4450-9fcb-b505a7a1d969"). InnerVolumeSpecName "kube-api-access-q6pk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.856852 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "76cb1e0e-4321-4450-9fcb-b505a7a1d969" (UID: "76cb1e0e-4321-4450-9fcb-b505a7a1d969"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.856986 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "76cb1e0e-4321-4450-9fcb-b505a7a1d969" (UID: "76cb1e0e-4321-4450-9fcb-b505a7a1d969"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.857143 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "76cb1e0e-4321-4450-9fcb-b505a7a1d969" (UID: "76cb1e0e-4321-4450-9fcb-b505a7a1d969"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.857493 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "76cb1e0e-4321-4450-9fcb-b505a7a1d969" (UID: "76cb1e0e-4321-4450-9fcb-b505a7a1d969"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.857709 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "76cb1e0e-4321-4450-9fcb-b505a7a1d969" (UID: "76cb1e0e-4321-4450-9fcb-b505a7a1d969"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.946709 4896 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.946735 4896 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.946746 4896 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.946758 4896 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.946766 4896 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.946776 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6pk7\" (UniqueName: \"kubernetes.io/projected/76cb1e0e-4321-4450-9fcb-b505a7a1d969-kube-api-access-q6pk7\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.946785 4896 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.946794 4896 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.946803 4896 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/76cb1e0e-4321-4450-9fcb-b505a7a1d969-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.946812 4896 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.946821 4896 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.946830 4896 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/76cb1e0e-4321-4450-9fcb-b505a7a1d969-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:19 crc kubenswrapper[4896]: I0218 14:59:19.946839 4896 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/76cb1e0e-4321-4450-9fcb-b505a7a1d969-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:20 crc kubenswrapper[4896]: I0218 14:59:20.323941 4896 generic.go:334] "Generic (PLEG): container finished" podID="76cb1e0e-4321-4450-9fcb-b505a7a1d969" containerID="fce8bf5bd3d671ffcdafbfabccda9c929d8b1b87b96a506cadf5a5c962411ee1" exitCode=0 Feb 18 14:59:20 crc kubenswrapper[4896]: I0218 14:59:20.323996 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" Feb 18 14:59:20 crc kubenswrapper[4896]: I0218 14:59:20.324010 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" event={"ID":"76cb1e0e-4321-4450-9fcb-b505a7a1d969","Type":"ContainerDied","Data":"fce8bf5bd3d671ffcdafbfabccda9c929d8b1b87b96a506cadf5a5c962411ee1"} Feb 18 14:59:20 crc kubenswrapper[4896]: I0218 14:59:20.324052 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-shvc6" event={"ID":"76cb1e0e-4321-4450-9fcb-b505a7a1d969","Type":"ContainerDied","Data":"a46c46add4951b8d35cc1f7cd2f6966369f1821f641e11c09321c0d99de9be14"} Feb 18 14:59:20 crc kubenswrapper[4896]: I0218 14:59:20.324074 4896 scope.go:117] "RemoveContainer" containerID="fce8bf5bd3d671ffcdafbfabccda9c929d8b1b87b96a506cadf5a5c962411ee1" Feb 18 14:59:20 crc kubenswrapper[4896]: I0218 14:59:20.343483 4896 scope.go:117] "RemoveContainer" containerID="fce8bf5bd3d671ffcdafbfabccda9c929d8b1b87b96a506cadf5a5c962411ee1" Feb 18 14:59:20 crc kubenswrapper[4896]: E0218 14:59:20.344138 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fce8bf5bd3d671ffcdafbfabccda9c929d8b1b87b96a506cadf5a5c962411ee1\": container with ID starting with fce8bf5bd3d671ffcdafbfabccda9c929d8b1b87b96a506cadf5a5c962411ee1 not found: ID does not exist" containerID="fce8bf5bd3d671ffcdafbfabccda9c929d8b1b87b96a506cadf5a5c962411ee1" Feb 18 14:59:20 crc kubenswrapper[4896]: I0218 14:59:20.344199 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fce8bf5bd3d671ffcdafbfabccda9c929d8b1b87b96a506cadf5a5c962411ee1"} err="failed to get container status \"fce8bf5bd3d671ffcdafbfabccda9c929d8b1b87b96a506cadf5a5c962411ee1\": rpc error: code = NotFound desc = could not find container \"fce8bf5bd3d671ffcdafbfabccda9c929d8b1b87b96a506cadf5a5c962411ee1\": container with ID starting with fce8bf5bd3d671ffcdafbfabccda9c929d8b1b87b96a506cadf5a5c962411ee1 not found: ID does not exist" Feb 18 14:59:24 crc kubenswrapper[4896]: I0218 14:59:24.500164 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 18 14:59:25 crc kubenswrapper[4896]: I0218 14:59:25.018543 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 18 14:59:25 crc kubenswrapper[4896]: I0218 14:59:25.587433 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 18 14:59:26 crc kubenswrapper[4896]: I0218 14:59:26.416085 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 18 14:59:28 crc kubenswrapper[4896]: I0218 14:59:28.256702 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 18 14:59:28 crc kubenswrapper[4896]: I0218 14:59:28.597983 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 18 14:59:28 crc kubenswrapper[4896]: I0218 14:59:28.861424 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 18 14:59:29 crc kubenswrapper[4896]: I0218 14:59:29.243155 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 18 14:59:29 crc kubenswrapper[4896]: I0218 14:59:29.335101 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 18 14:59:29 crc kubenswrapper[4896]: I0218 14:59:29.344486 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 18 14:59:29 crc kubenswrapper[4896]: I0218 14:59:29.419436 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 18 14:59:29 crc kubenswrapper[4896]: I0218 14:59:29.588078 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 18 14:59:29 crc kubenswrapper[4896]: I0218 14:59:29.616361 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 18 14:59:29 crc kubenswrapper[4896]: I0218 14:59:29.709815 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 18 14:59:29 crc kubenswrapper[4896]: I0218 14:59:29.725683 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 18 14:59:29 crc kubenswrapper[4896]: I0218 14:59:29.828177 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 18 14:59:29 crc kubenswrapper[4896]: I0218 14:59:29.982135 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 18 14:59:30 crc kubenswrapper[4896]: I0218 14:59:30.150507 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 18 14:59:30 crc kubenswrapper[4896]: I0218 14:59:30.154080 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 18 14:59:30 crc kubenswrapper[4896]: I0218 14:59:30.582191 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 18 14:59:30 crc kubenswrapper[4896]: I0218 14:59:30.726946 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 18 14:59:30 crc kubenswrapper[4896]: I0218 14:59:30.781364 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 18 14:59:30 crc kubenswrapper[4896]: I0218 14:59:30.784928 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 18 14:59:30 crc kubenswrapper[4896]: I0218 14:59:30.889130 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 18 14:59:30 crc kubenswrapper[4896]: I0218 14:59:30.922107 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 18 14:59:31 crc kubenswrapper[4896]: I0218 14:59:31.043651 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 18 14:59:31 crc kubenswrapper[4896]: I0218 14:59:31.587588 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 18 14:59:31 crc kubenswrapper[4896]: I0218 14:59:31.662623 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 18 14:59:31 crc kubenswrapper[4896]: I0218 14:59:31.762131 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 18 14:59:31 crc kubenswrapper[4896]: I0218 14:59:31.884513 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 18 14:59:31 crc kubenswrapper[4896]: I0218 14:59:31.915607 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 18 14:59:31 crc kubenswrapper[4896]: I0218 14:59:31.941093 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 18 14:59:31 crc kubenswrapper[4896]: I0218 14:59:31.954197 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 18 14:59:31 crc kubenswrapper[4896]: I0218 14:59:31.965450 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 18 14:59:31 crc kubenswrapper[4896]: I0218 14:59:31.973078 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 18 14:59:32 crc kubenswrapper[4896]: I0218 14:59:32.007014 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 18 14:59:32 crc kubenswrapper[4896]: I0218 14:59:32.090709 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 18 14:59:32 crc kubenswrapper[4896]: I0218 14:59:32.503397 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 18 14:59:32 crc kubenswrapper[4896]: I0218 14:59:32.613319 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 18 14:59:32 crc kubenswrapper[4896]: I0218 14:59:32.649522 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 18 14:59:32 crc kubenswrapper[4896]: I0218 14:59:32.667580 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 18 14:59:32 crc kubenswrapper[4896]: I0218 14:59:32.781341 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 18 14:59:32 crc kubenswrapper[4896]: I0218 14:59:32.884084 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 18 14:59:32 crc kubenswrapper[4896]: I0218 14:59:32.902307 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 18 14:59:33 crc kubenswrapper[4896]: I0218 14:59:33.167151 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 18 14:59:33 crc kubenswrapper[4896]: I0218 14:59:33.208821 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 18 14:59:33 crc kubenswrapper[4896]: I0218 14:59:33.227704 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 18 14:59:33 crc kubenswrapper[4896]: I0218 14:59:33.403535 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 18 14:59:33 crc kubenswrapper[4896]: I0218 14:59:33.466205 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 18 14:59:33 crc kubenswrapper[4896]: I0218 14:59:33.538900 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 18 14:59:33 crc kubenswrapper[4896]: I0218 14:59:33.690103 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 18 14:59:33 crc kubenswrapper[4896]: I0218 14:59:33.929837 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 18 14:59:34 crc kubenswrapper[4896]: I0218 14:59:34.095922 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 18 14:59:34 crc kubenswrapper[4896]: I0218 14:59:34.101046 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 18 14:59:34 crc kubenswrapper[4896]: I0218 14:59:34.242791 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 18 14:59:34 crc kubenswrapper[4896]: I0218 14:59:34.255001 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 18 14:59:34 crc kubenswrapper[4896]: I0218 14:59:34.307685 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 18 14:59:34 crc kubenswrapper[4896]: I0218 14:59:34.384832 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 18 14:59:34 crc kubenswrapper[4896]: I0218 14:59:34.488791 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 18 14:59:34 crc kubenswrapper[4896]: I0218 14:59:34.684207 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 18 14:59:34 crc kubenswrapper[4896]: I0218 14:59:34.757411 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 18 14:59:34 crc kubenswrapper[4896]: I0218 14:59:34.775910 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 18 14:59:34 crc kubenswrapper[4896]: I0218 14:59:34.848457 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 18 14:59:34 crc kubenswrapper[4896]: I0218 14:59:34.861480 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 18 14:59:34 crc kubenswrapper[4896]: I0218 14:59:34.917787 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 18 14:59:34 crc kubenswrapper[4896]: I0218 14:59:34.925818 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 18 14:59:34 crc kubenswrapper[4896]: I0218 14:59:34.967369 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 18 14:59:35 crc kubenswrapper[4896]: I0218 14:59:35.177463 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 18 14:59:35 crc kubenswrapper[4896]: I0218 14:59:35.210169 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 18 14:59:35 crc kubenswrapper[4896]: I0218 14:59:35.279980 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 18 14:59:35 crc kubenswrapper[4896]: I0218 14:59:35.427092 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 18 14:59:35 crc kubenswrapper[4896]: I0218 14:59:35.531669 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 18 14:59:35 crc kubenswrapper[4896]: I0218 14:59:35.591557 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 18 14:59:35 crc kubenswrapper[4896]: I0218 14:59:35.636637 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 18 14:59:35 crc kubenswrapper[4896]: I0218 14:59:35.638030 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 18 14:59:35 crc kubenswrapper[4896]: I0218 14:59:35.681409 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 18 14:59:35 crc kubenswrapper[4896]: I0218 14:59:35.790478 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 18 14:59:35 crc kubenswrapper[4896]: I0218 14:59:35.844737 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 18 14:59:35 crc kubenswrapper[4896]: I0218 14:59:35.865195 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 18 14:59:35 crc kubenswrapper[4896]: I0218 14:59:35.892253 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 18 14:59:36 crc kubenswrapper[4896]: I0218 14:59:36.109811 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 18 14:59:36 crc kubenswrapper[4896]: I0218 14:59:36.129762 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 18 14:59:36 crc kubenswrapper[4896]: I0218 14:59:36.132867 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 18 14:59:36 crc kubenswrapper[4896]: I0218 14:59:36.247649 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 18 14:59:36 crc kubenswrapper[4896]: I0218 14:59:36.252465 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 18 14:59:36 crc kubenswrapper[4896]: I0218 14:59:36.314452 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 18 14:59:36 crc kubenswrapper[4896]: I0218 14:59:36.354954 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 18 14:59:36 crc kubenswrapper[4896]: I0218 14:59:36.380940 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 18 14:59:36 crc kubenswrapper[4896]: I0218 14:59:36.389186 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 18 14:59:36 crc kubenswrapper[4896]: I0218 14:59:36.588000 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 18 14:59:36 crc kubenswrapper[4896]: I0218 14:59:36.616829 4896 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 18 14:59:36 crc kubenswrapper[4896]: I0218 14:59:36.617222 4896 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 18 14:59:36 crc kubenswrapper[4896]: I0218 14:59:36.743387 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 18 14:59:36 crc kubenswrapper[4896]: I0218 14:59:36.910126 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 18 14:59:36 crc kubenswrapper[4896]: I0218 14:59:36.956714 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 18 14:59:36 crc kubenswrapper[4896]: I0218 14:59:36.959729 4896 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 18 14:59:36 crc kubenswrapper[4896]: I0218 14:59:36.965524 4896 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-shvc6","openshift-kube-apiserver/kube-apiserver-crc","openshift-marketplace/redhat-operators-t6sfb"] Feb 18 14:59:36 crc kubenswrapper[4896]: I0218 14:59:36.965605 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 18 14:59:36 crc kubenswrapper[4896]: I0218 14:59:36.973742 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:59:36 crc kubenswrapper[4896]: I0218 14:59:36.986022 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=19.986001881 podStartE2EDuration="19.986001881s" podCreationTimestamp="2026-02-18 14:59:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:59:36.984981143 +0000 UTC m=+275.914315356" watchObservedRunningTime="2026-02-18 14:59:36.986001881 +0000 UTC m=+275.915336104" Feb 18 14:59:37 crc kubenswrapper[4896]: I0218 14:59:37.009656 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 18 14:59:37 crc kubenswrapper[4896]: I0218 14:59:37.108504 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 18 14:59:37 crc kubenswrapper[4896]: I0218 14:59:37.128495 4896 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 18 14:59:37 crc kubenswrapper[4896]: I0218 14:59:37.186515 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 18 14:59:37 crc kubenswrapper[4896]: I0218 14:59:37.259102 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 18 14:59:37 crc kubenswrapper[4896]: I0218 14:59:37.307560 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 18 14:59:37 crc kubenswrapper[4896]: I0218 14:59:37.351294 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 18 14:59:37 crc kubenswrapper[4896]: I0218 14:59:37.354628 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 18 14:59:37 crc kubenswrapper[4896]: I0218 14:59:37.400736 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 18 14:59:37 crc kubenswrapper[4896]: I0218 14:59:37.411704 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 18 14:59:37 crc kubenswrapper[4896]: I0218 14:59:37.416619 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 18 14:59:37 crc kubenswrapper[4896]: I0218 14:59:37.423007 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 18 14:59:37 crc kubenswrapper[4896]: I0218 14:59:37.461362 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 18 14:59:37 crc kubenswrapper[4896]: I0218 14:59:37.745564 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 18 14:59:37 crc kubenswrapper[4896]: I0218 14:59:37.845004 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 18 14:59:37 crc kubenswrapper[4896]: I0218 14:59:37.861110 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 18 14:59:37 crc kubenswrapper[4896]: I0218 14:59:37.900299 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 18 14:59:37 crc kubenswrapper[4896]: I0218 14:59:37.937123 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76cb1e0e-4321-4450-9fcb-b505a7a1d969" path="/var/lib/kubelet/pods/76cb1e0e-4321-4450-9fcb-b505a7a1d969/volumes" Feb 18 14:59:37 crc kubenswrapper[4896]: I0218 14:59:37.938043 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eddbcdff-a387-4f95-9277-5850b6485036" path="/var/lib/kubelet/pods/eddbcdff-a387-4f95-9277-5850b6485036/volumes" Feb 18 14:59:37 crc kubenswrapper[4896]: I0218 14:59:37.960520 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 18 14:59:37 crc kubenswrapper[4896]: I0218 14:59:37.971454 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 18 14:59:38 crc kubenswrapper[4896]: I0218 14:59:38.031369 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 18 14:59:38 crc kubenswrapper[4896]: I0218 14:59:38.037654 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 18 14:59:38 crc kubenswrapper[4896]: I0218 14:59:38.057078 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 18 14:59:38 crc kubenswrapper[4896]: I0218 14:59:38.082263 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 18 14:59:38 crc kubenswrapper[4896]: I0218 14:59:38.118358 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 18 14:59:38 crc kubenswrapper[4896]: I0218 14:59:38.141873 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 18 14:59:38 crc kubenswrapper[4896]: I0218 14:59:38.210101 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 18 14:59:38 crc kubenswrapper[4896]: I0218 14:59:38.303383 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 18 14:59:38 crc kubenswrapper[4896]: I0218 14:59:38.318890 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 18 14:59:38 crc kubenswrapper[4896]: I0218 14:59:38.348001 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 18 14:59:38 crc kubenswrapper[4896]: I0218 14:59:38.457299 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 18 14:59:38 crc kubenswrapper[4896]: I0218 14:59:38.486661 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 18 14:59:38 crc kubenswrapper[4896]: I0218 14:59:38.584102 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 18 14:59:38 crc kubenswrapper[4896]: I0218 14:59:38.646892 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 18 14:59:38 crc kubenswrapper[4896]: I0218 14:59:38.664631 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 18 14:59:38 crc kubenswrapper[4896]: I0218 14:59:38.687887 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 18 14:59:38 crc kubenswrapper[4896]: I0218 14:59:38.723328 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 18 14:59:38 crc kubenswrapper[4896]: I0218 14:59:38.817542 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 18 14:59:38 crc kubenswrapper[4896]: I0218 14:59:38.949443 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 18 14:59:38 crc kubenswrapper[4896]: I0218 14:59:38.966098 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 18 14:59:38 crc kubenswrapper[4896]: I0218 14:59:38.988340 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 18 14:59:39 crc kubenswrapper[4896]: I0218 14:59:38.999975 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 18 14:59:39 crc kubenswrapper[4896]: I0218 14:59:39.049738 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 18 14:59:39 crc kubenswrapper[4896]: I0218 14:59:39.053977 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 18 14:59:39 crc kubenswrapper[4896]: I0218 14:59:39.105524 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 18 14:59:39 crc kubenswrapper[4896]: I0218 14:59:39.176248 4896 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 18 14:59:39 crc kubenswrapper[4896]: I0218 14:59:39.176507 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://df25631ac2f71f0842be433055c01f20cb4d72be8041d8d32c5ff0b4cdd6866c" gracePeriod=5 Feb 18 14:59:39 crc kubenswrapper[4896]: I0218 14:59:39.200533 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 18 14:59:39 crc kubenswrapper[4896]: I0218 14:59:39.221358 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 18 14:59:39 crc kubenswrapper[4896]: I0218 14:59:39.378332 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 18 14:59:39 crc kubenswrapper[4896]: I0218 14:59:39.476928 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 18 14:59:39 crc kubenswrapper[4896]: I0218 14:59:39.629971 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 18 14:59:39 crc kubenswrapper[4896]: I0218 14:59:39.675843 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 18 14:59:39 crc kubenswrapper[4896]: I0218 14:59:39.688612 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 18 14:59:39 crc kubenswrapper[4896]: I0218 14:59:39.842121 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 18 14:59:39 crc kubenswrapper[4896]: I0218 14:59:39.899041 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 18 14:59:39 crc kubenswrapper[4896]: I0218 14:59:39.922245 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 18 14:59:40 crc kubenswrapper[4896]: I0218 14:59:40.027840 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 18 14:59:40 crc kubenswrapper[4896]: I0218 14:59:40.079296 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 18 14:59:40 crc kubenswrapper[4896]: I0218 14:59:40.148529 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 18 14:59:40 crc kubenswrapper[4896]: I0218 14:59:40.345346 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 18 14:59:40 crc kubenswrapper[4896]: I0218 14:59:40.427684 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 18 14:59:40 crc kubenswrapper[4896]: I0218 14:59:40.472862 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 18 14:59:40 crc kubenswrapper[4896]: I0218 14:59:40.487287 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 18 14:59:40 crc kubenswrapper[4896]: I0218 14:59:40.757327 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 18 14:59:40 crc kubenswrapper[4896]: I0218 14:59:40.846028 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 18 14:59:40 crc kubenswrapper[4896]: I0218 14:59:40.873713 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 18 14:59:40 crc kubenswrapper[4896]: I0218 14:59:40.969117 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 18 14:59:41 crc kubenswrapper[4896]: I0218 14:59:41.014037 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 18 14:59:41 crc kubenswrapper[4896]: I0218 14:59:41.103699 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 18 14:59:41 crc kubenswrapper[4896]: I0218 14:59:41.112514 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 18 14:59:41 crc kubenswrapper[4896]: I0218 14:59:41.159570 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 18 14:59:41 crc kubenswrapper[4896]: I0218 14:59:41.177032 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 18 14:59:41 crc kubenswrapper[4896]: I0218 14:59:41.200725 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 18 14:59:41 crc kubenswrapper[4896]: I0218 14:59:41.209688 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 18 14:59:41 crc kubenswrapper[4896]: I0218 14:59:41.227080 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 18 14:59:41 crc kubenswrapper[4896]: I0218 14:59:41.236041 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 18 14:59:41 crc kubenswrapper[4896]: I0218 14:59:41.263750 4896 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 18 14:59:41 crc kubenswrapper[4896]: I0218 14:59:41.266071 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 18 14:59:41 crc kubenswrapper[4896]: I0218 14:59:41.296410 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 18 14:59:41 crc kubenswrapper[4896]: I0218 14:59:41.459448 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 18 14:59:41 crc kubenswrapper[4896]: I0218 14:59:41.578764 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 18 14:59:41 crc kubenswrapper[4896]: I0218 14:59:41.735661 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 18 14:59:41 crc kubenswrapper[4896]: I0218 14:59:41.743976 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 18 14:59:41 crc kubenswrapper[4896]: I0218 14:59:41.762729 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 18 14:59:41 crc kubenswrapper[4896]: I0218 14:59:41.818139 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 18 14:59:41 crc kubenswrapper[4896]: I0218 14:59:41.845128 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 18 14:59:41 crc kubenswrapper[4896]: I0218 14:59:41.872224 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 18 14:59:41 crc kubenswrapper[4896]: I0218 14:59:41.913283 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.015155 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.051191 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.100004 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.125107 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.132724 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.217497 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.247431 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.321667 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-dc8679f5f-z9fj6"] Feb 18 14:59:42 crc kubenswrapper[4896]: E0218 14:59:42.322263 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76cb1e0e-4321-4450-9fcb-b505a7a1d969" containerName="oauth-openshift" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.322280 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="76cb1e0e-4321-4450-9fcb-b505a7a1d969" containerName="oauth-openshift" Feb 18 14:59:42 crc kubenswrapper[4896]: E0218 14:59:42.322293 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a847ad6-23c1-43e2-b916-c97df45c2aff" containerName="installer" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.322299 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a847ad6-23c1-43e2-b916-c97df45c2aff" containerName="installer" Feb 18 14:59:42 crc kubenswrapper[4896]: E0218 14:59:42.322308 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eddbcdff-a387-4f95-9277-5850b6485036" containerName="extract-utilities" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.322314 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="eddbcdff-a387-4f95-9277-5850b6485036" containerName="extract-utilities" Feb 18 14:59:42 crc kubenswrapper[4896]: E0218 14:59:42.322325 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.322330 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 18 14:59:42 crc kubenswrapper[4896]: E0218 14:59:42.322342 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eddbcdff-a387-4f95-9277-5850b6485036" containerName="registry-server" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.322347 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="eddbcdff-a387-4f95-9277-5850b6485036" containerName="registry-server" Feb 18 14:59:42 crc kubenswrapper[4896]: E0218 14:59:42.322355 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eddbcdff-a387-4f95-9277-5850b6485036" containerName="extract-content" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.322360 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="eddbcdff-a387-4f95-9277-5850b6485036" containerName="extract-content" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.322469 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="eddbcdff-a387-4f95-9277-5850b6485036" containerName="registry-server" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.322482 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.322489 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a847ad6-23c1-43e2-b916-c97df45c2aff" containerName="installer" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.322500 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="76cb1e0e-4321-4450-9fcb-b505a7a1d969" containerName="oauth-openshift" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.322823 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.329518 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.329782 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.329976 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.330041 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.330046 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.329968 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.330147 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.330270 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.330533 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.330925 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.331369 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.333571 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.343118 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.345464 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.345921 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-dc8679f5f-z9fj6"] Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.352515 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.411903 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.426332 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/62a98176-51f2-4270-882b-d10d827117a4-v4-0-config-user-template-error\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.426500 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/62a98176-51f2-4270-882b-d10d827117a4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.426671 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6v7r\" (UniqueName: \"kubernetes.io/projected/62a98176-51f2-4270-882b-d10d827117a4-kube-api-access-j6v7r\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.427189 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/62a98176-51f2-4270-882b-d10d827117a4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.427277 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/62a98176-51f2-4270-882b-d10d827117a4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.427318 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/62a98176-51f2-4270-882b-d10d827117a4-v4-0-config-system-router-certs\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.427351 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/62a98176-51f2-4270-882b-d10d827117a4-v4-0-config-user-template-login\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.427468 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/62a98176-51f2-4270-882b-d10d827117a4-v4-0-config-system-session\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.427547 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/62a98176-51f2-4270-882b-d10d827117a4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.427642 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/62a98176-51f2-4270-882b-d10d827117a4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.427690 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/62a98176-51f2-4270-882b-d10d827117a4-v4-0-config-system-service-ca\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.427885 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/62a98176-51f2-4270-882b-d10d827117a4-audit-policies\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.428050 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/62a98176-51f2-4270-882b-d10d827117a4-audit-dir\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.428114 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/62a98176-51f2-4270-882b-d10d827117a4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.428171 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.491001 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.524056 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.529379 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/62a98176-51f2-4270-882b-d10d827117a4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.529409 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/62a98176-51f2-4270-882b-d10d827117a4-v4-0-config-system-service-ca\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.529438 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/62a98176-51f2-4270-882b-d10d827117a4-audit-policies\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.529461 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/62a98176-51f2-4270-882b-d10d827117a4-audit-dir\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.529479 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/62a98176-51f2-4270-882b-d10d827117a4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.529503 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/62a98176-51f2-4270-882b-d10d827117a4-v4-0-config-user-template-error\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.529522 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/62a98176-51f2-4270-882b-d10d827117a4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.529550 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6v7r\" (UniqueName: \"kubernetes.io/projected/62a98176-51f2-4270-882b-d10d827117a4-kube-api-access-j6v7r\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.529571 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/62a98176-51f2-4270-882b-d10d827117a4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.529586 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/62a98176-51f2-4270-882b-d10d827117a4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.529606 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/62a98176-51f2-4270-882b-d10d827117a4-v4-0-config-user-template-login\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.529626 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/62a98176-51f2-4270-882b-d10d827117a4-v4-0-config-system-router-certs\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.529645 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/62a98176-51f2-4270-882b-d10d827117a4-v4-0-config-system-session\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.529663 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/62a98176-51f2-4270-882b-d10d827117a4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.529897 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/62a98176-51f2-4270-882b-d10d827117a4-audit-dir\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.530537 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/62a98176-51f2-4270-882b-d10d827117a4-v4-0-config-system-service-ca\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.531064 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/62a98176-51f2-4270-882b-d10d827117a4-audit-policies\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.531176 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/62a98176-51f2-4270-882b-d10d827117a4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.533595 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/62a98176-51f2-4270-882b-d10d827117a4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.535185 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/62a98176-51f2-4270-882b-d10d827117a4-v4-0-config-user-template-error\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.536472 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/62a98176-51f2-4270-882b-d10d827117a4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.536513 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/62a98176-51f2-4270-882b-d10d827117a4-v4-0-config-system-router-certs\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.537408 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/62a98176-51f2-4270-882b-d10d827117a4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.537531 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/62a98176-51f2-4270-882b-d10d827117a4-v4-0-config-system-session\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.538332 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/62a98176-51f2-4270-882b-d10d827117a4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.542021 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/62a98176-51f2-4270-882b-d10d827117a4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.543737 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/62a98176-51f2-4270-882b-d10d827117a4-v4-0-config-user-template-login\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.548877 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6v7r\" (UniqueName: \"kubernetes.io/projected/62a98176-51f2-4270-882b-d10d827117a4-kube-api-access-j6v7r\") pod \"oauth-openshift-dc8679f5f-z9fj6\" (UID: \"62a98176-51f2-4270-882b-d10d827117a4\") " pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.615437 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.644416 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.663061 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.668173 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.695658 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.868869 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.895794 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 18 14:59:42 crc kubenswrapper[4896]: I0218 14:59:42.962446 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 18 14:59:43 crc kubenswrapper[4896]: I0218 14:59:43.010630 4896 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 18 14:59:43 crc kubenswrapper[4896]: I0218 14:59:43.041787 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 18 14:59:43 crc kubenswrapper[4896]: I0218 14:59:43.213534 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 18 14:59:43 crc kubenswrapper[4896]: I0218 14:59:43.214479 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-dc8679f5f-z9fj6"] Feb 18 14:59:43 crc kubenswrapper[4896]: I0218 14:59:43.238006 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 18 14:59:43 crc kubenswrapper[4896]: I0218 14:59:43.367941 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 18 14:59:43 crc kubenswrapper[4896]: I0218 14:59:43.443544 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" event={"ID":"62a98176-51f2-4270-882b-d10d827117a4","Type":"ContainerStarted","Data":"997af11bfd2e1bcf5e7dc3c7e31f9101799325c617d2b0b582f7cd59e20d2056"} Feb 18 14:59:43 crc kubenswrapper[4896]: I0218 14:59:43.474658 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 18 14:59:43 crc kubenswrapper[4896]: I0218 14:59:43.488894 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 18 14:59:43 crc kubenswrapper[4896]: I0218 14:59:43.638225 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 18 14:59:43 crc kubenswrapper[4896]: I0218 14:59:43.771100 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 18 14:59:43 crc kubenswrapper[4896]: I0218 14:59:43.799370 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 18 14:59:43 crc kubenswrapper[4896]: I0218 14:59:43.861710 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 18 14:59:43 crc kubenswrapper[4896]: I0218 14:59:43.884077 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 18 14:59:43 crc kubenswrapper[4896]: I0218 14:59:43.936480 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 18 14:59:43 crc kubenswrapper[4896]: I0218 14:59:43.954026 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 18 14:59:43 crc kubenswrapper[4896]: I0218 14:59:43.991999 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.104681 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.137420 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.152065 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.204040 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 18 14:59:44 crc kubenswrapper[4896]: E0218 14:59:44.280435 4896 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-conmon-df25631ac2f71f0842be433055c01f20cb4d72be8041d8d32c5ff0b4cdd6866c.scope\": RecentStats: unable to find data in memory cache]" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.289265 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.289328 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.351270 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.351634 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.351634 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.351690 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.351706 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.351721 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.351744 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.351783 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.351849 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.352089 4896 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.352111 4896 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.352122 4896 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.352133 4896 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.359281 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.371704 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.403768 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.449597 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" event={"ID":"62a98176-51f2-4270-882b-d10d827117a4","Type":"ContainerStarted","Data":"855782dfeea150133cb211b11a24ae49703ad814cbc3f373b9274748edbb8488"} Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.450149 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.452648 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.452722 4896 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="df25631ac2f71f0842be433055c01f20cb4d72be8041d8d32c5ff0b4cdd6866c" exitCode=137 Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.452759 4896 scope.go:117] "RemoveContainer" containerID="df25631ac2f71f0842be433055c01f20cb4d72be8041d8d32c5ff0b4cdd6866c" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.452888 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.452991 4896 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.457582 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.468326 4896 scope.go:117] "RemoveContainer" containerID="df25631ac2f71f0842be433055c01f20cb4d72be8041d8d32c5ff0b4cdd6866c" Feb 18 14:59:44 crc kubenswrapper[4896]: E0218 14:59:44.468987 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df25631ac2f71f0842be433055c01f20cb4d72be8041d8d32c5ff0b4cdd6866c\": container with ID starting with df25631ac2f71f0842be433055c01f20cb4d72be8041d8d32c5ff0b4cdd6866c not found: ID does not exist" containerID="df25631ac2f71f0842be433055c01f20cb4d72be8041d8d32c5ff0b4cdd6866c" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.469033 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df25631ac2f71f0842be433055c01f20cb4d72be8041d8d32c5ff0b4cdd6866c"} err="failed to get container status \"df25631ac2f71f0842be433055c01f20cb4d72be8041d8d32c5ff0b4cdd6866c\": rpc error: code = NotFound desc = could not find container \"df25631ac2f71f0842be433055c01f20cb4d72be8041d8d32c5ff0b4cdd6866c\": container with ID starting with df25631ac2f71f0842be433055c01f20cb4d72be8041d8d32c5ff0b4cdd6866c not found: ID does not exist" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.496484 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-dc8679f5f-z9fj6" podStartSLOduration=50.496459851 podStartE2EDuration="50.496459851s" podCreationTimestamp="2026-02-18 14:58:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:59:44.468597653 +0000 UTC m=+283.397931866" watchObservedRunningTime="2026-02-18 14:59:44.496459851 +0000 UTC m=+283.425794104" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.500612 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.512539 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.629675 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.663380 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.697568 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.828981 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.885780 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 18 14:59:44 crc kubenswrapper[4896]: I0218 14:59:44.947839 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 18 14:59:45 crc kubenswrapper[4896]: I0218 14:59:45.271297 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 18 14:59:45 crc kubenswrapper[4896]: I0218 14:59:45.331804 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 18 14:59:45 crc kubenswrapper[4896]: I0218 14:59:45.409634 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 18 14:59:45 crc kubenswrapper[4896]: I0218 14:59:45.418110 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 18 14:59:45 crc kubenswrapper[4896]: I0218 14:59:45.508145 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 18 14:59:45 crc kubenswrapper[4896]: I0218 14:59:45.850660 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 18 14:59:45 crc kubenswrapper[4896]: I0218 14:59:45.935790 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Feb 18 14:59:45 crc kubenswrapper[4896]: I0218 14:59:45.988030 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 18 14:59:46 crc kubenswrapper[4896]: I0218 14:59:46.328647 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 18 14:59:47 crc kubenswrapper[4896]: I0218 14:59:47.534899 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 18 14:59:47 crc kubenswrapper[4896]: I0218 14:59:47.652529 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 18 14:59:50 crc kubenswrapper[4896]: I0218 14:59:50.521721 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-c6db4b444-cr7c9"] Feb 18 14:59:50 crc kubenswrapper[4896]: I0218 14:59:50.522352 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-c6db4b444-cr7c9" podUID="e2590801-a8b2-442a-af31-b03b63138aa3" containerName="controller-manager" containerID="cri-o://2f1fde29ede1d3cd3fe1ce2d38113ba249cd5ba85e7da9b6748379d9eba796fb" gracePeriod=30 Feb 18 14:59:50 crc kubenswrapper[4896]: I0218 14:59:50.627051 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c887bd576-rw68b"] Feb 18 14:59:50 crc kubenswrapper[4896]: I0218 14:59:50.628420 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-c887bd576-rw68b" podUID="85c1cb1a-6343-490c-98a3-f11a52290a73" containerName="route-controller-manager" containerID="cri-o://de9e0424956f2a0744091ce1d08afeeedd630f291f0e017c8cc1ad4e08ea4db4" gracePeriod=30 Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.483868 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-c6db4b444-cr7c9" Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.489494 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c887bd576-rw68b" Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.491290 4896 generic.go:334] "Generic (PLEG): container finished" podID="e2590801-a8b2-442a-af31-b03b63138aa3" containerID="2f1fde29ede1d3cd3fe1ce2d38113ba249cd5ba85e7da9b6748379d9eba796fb" exitCode=0 Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.491362 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-c6db4b444-cr7c9" Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.491561 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-c6db4b444-cr7c9" event={"ID":"e2590801-a8b2-442a-af31-b03b63138aa3","Type":"ContainerDied","Data":"2f1fde29ede1d3cd3fe1ce2d38113ba249cd5ba85e7da9b6748379d9eba796fb"} Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.491589 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-c6db4b444-cr7c9" event={"ID":"e2590801-a8b2-442a-af31-b03b63138aa3","Type":"ContainerDied","Data":"2663fae556f2ab06c94adf2594e4112106632266d939fa459b02970aaad80d66"} Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.491605 4896 scope.go:117] "RemoveContainer" containerID="2f1fde29ede1d3cd3fe1ce2d38113ba249cd5ba85e7da9b6748379d9eba796fb" Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.492870 4896 generic.go:334] "Generic (PLEG): container finished" podID="85c1cb1a-6343-490c-98a3-f11a52290a73" containerID="de9e0424956f2a0744091ce1d08afeeedd630f291f0e017c8cc1ad4e08ea4db4" exitCode=0 Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.492890 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c887bd576-rw68b" event={"ID":"85c1cb1a-6343-490c-98a3-f11a52290a73","Type":"ContainerDied","Data":"de9e0424956f2a0744091ce1d08afeeedd630f291f0e017c8cc1ad4e08ea4db4"} Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.492901 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c887bd576-rw68b" event={"ID":"85c1cb1a-6343-490c-98a3-f11a52290a73","Type":"ContainerDied","Data":"94e81e500693312ac13759e34f54082033edb562b8b9e75959b6c382f08bb8a5"} Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.492929 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c887bd576-rw68b" Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.514326 4896 scope.go:117] "RemoveContainer" containerID="2f1fde29ede1d3cd3fe1ce2d38113ba249cd5ba85e7da9b6748379d9eba796fb" Feb 18 14:59:51 crc kubenswrapper[4896]: E0218 14:59:51.514935 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f1fde29ede1d3cd3fe1ce2d38113ba249cd5ba85e7da9b6748379d9eba796fb\": container with ID starting with 2f1fde29ede1d3cd3fe1ce2d38113ba249cd5ba85e7da9b6748379d9eba796fb not found: ID does not exist" containerID="2f1fde29ede1d3cd3fe1ce2d38113ba249cd5ba85e7da9b6748379d9eba796fb" Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.517353 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f1fde29ede1d3cd3fe1ce2d38113ba249cd5ba85e7da9b6748379d9eba796fb"} err="failed to get container status \"2f1fde29ede1d3cd3fe1ce2d38113ba249cd5ba85e7da9b6748379d9eba796fb\": rpc error: code = NotFound desc = could not find container \"2f1fde29ede1d3cd3fe1ce2d38113ba249cd5ba85e7da9b6748379d9eba796fb\": container with ID starting with 2f1fde29ede1d3cd3fe1ce2d38113ba249cd5ba85e7da9b6748379d9eba796fb not found: ID does not exist" Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.517515 4896 scope.go:117] "RemoveContainer" containerID="de9e0424956f2a0744091ce1d08afeeedd630f291f0e017c8cc1ad4e08ea4db4" Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.538554 4896 scope.go:117] "RemoveContainer" containerID="de9e0424956f2a0744091ce1d08afeeedd630f291f0e017c8cc1ad4e08ea4db4" Feb 18 14:59:51 crc kubenswrapper[4896]: E0218 14:59:51.539663 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de9e0424956f2a0744091ce1d08afeeedd630f291f0e017c8cc1ad4e08ea4db4\": container with ID starting with de9e0424956f2a0744091ce1d08afeeedd630f291f0e017c8cc1ad4e08ea4db4 not found: ID does not exist" containerID="de9e0424956f2a0744091ce1d08afeeedd630f291f0e017c8cc1ad4e08ea4db4" Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.539688 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de9e0424956f2a0744091ce1d08afeeedd630f291f0e017c8cc1ad4e08ea4db4"} err="failed to get container status \"de9e0424956f2a0744091ce1d08afeeedd630f291f0e017c8cc1ad4e08ea4db4\": rpc error: code = NotFound desc = could not find container \"de9e0424956f2a0744091ce1d08afeeedd630f291f0e017c8cc1ad4e08ea4db4\": container with ID starting with de9e0424956f2a0744091ce1d08afeeedd630f291f0e017c8cc1ad4e08ea4db4 not found: ID does not exist" Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.634694 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/85c1cb1a-6343-490c-98a3-f11a52290a73-client-ca\") pod \"85c1cb1a-6343-490c-98a3-f11a52290a73\" (UID: \"85c1cb1a-6343-490c-98a3-f11a52290a73\") " Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.634748 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85c1cb1a-6343-490c-98a3-f11a52290a73-serving-cert\") pod \"85c1cb1a-6343-490c-98a3-f11a52290a73\" (UID: \"85c1cb1a-6343-490c-98a3-f11a52290a73\") " Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.634768 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85c1cb1a-6343-490c-98a3-f11a52290a73-config\") pod \"85c1cb1a-6343-490c-98a3-f11a52290a73\" (UID: \"85c1cb1a-6343-490c-98a3-f11a52290a73\") " Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.634786 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2590801-a8b2-442a-af31-b03b63138aa3-config\") pod \"e2590801-a8b2-442a-af31-b03b63138aa3\" (UID: \"e2590801-a8b2-442a-af31-b03b63138aa3\") " Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.634811 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e2590801-a8b2-442a-af31-b03b63138aa3-client-ca\") pod \"e2590801-a8b2-442a-af31-b03b63138aa3\" (UID: \"e2590801-a8b2-442a-af31-b03b63138aa3\") " Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.634831 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e2590801-a8b2-442a-af31-b03b63138aa3-serving-cert\") pod \"e2590801-a8b2-442a-af31-b03b63138aa3\" (UID: \"e2590801-a8b2-442a-af31-b03b63138aa3\") " Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.634865 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2xhl\" (UniqueName: \"kubernetes.io/projected/e2590801-a8b2-442a-af31-b03b63138aa3-kube-api-access-m2xhl\") pod \"e2590801-a8b2-442a-af31-b03b63138aa3\" (UID: \"e2590801-a8b2-442a-af31-b03b63138aa3\") " Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.634891 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49t2v\" (UniqueName: \"kubernetes.io/projected/85c1cb1a-6343-490c-98a3-f11a52290a73-kube-api-access-49t2v\") pod \"85c1cb1a-6343-490c-98a3-f11a52290a73\" (UID: \"85c1cb1a-6343-490c-98a3-f11a52290a73\") " Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.634908 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e2590801-a8b2-442a-af31-b03b63138aa3-proxy-ca-bundles\") pod \"e2590801-a8b2-442a-af31-b03b63138aa3\" (UID: \"e2590801-a8b2-442a-af31-b03b63138aa3\") " Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.636027 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85c1cb1a-6343-490c-98a3-f11a52290a73-client-ca" (OuterVolumeSpecName: "client-ca") pod "85c1cb1a-6343-490c-98a3-f11a52290a73" (UID: "85c1cb1a-6343-490c-98a3-f11a52290a73"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.637138 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2590801-a8b2-442a-af31-b03b63138aa3-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "e2590801-a8b2-442a-af31-b03b63138aa3" (UID: "e2590801-a8b2-442a-af31-b03b63138aa3"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.637302 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2590801-a8b2-442a-af31-b03b63138aa3-client-ca" (OuterVolumeSpecName: "client-ca") pod "e2590801-a8b2-442a-af31-b03b63138aa3" (UID: "e2590801-a8b2-442a-af31-b03b63138aa3"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.637331 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2590801-a8b2-442a-af31-b03b63138aa3-config" (OuterVolumeSpecName: "config") pod "e2590801-a8b2-442a-af31-b03b63138aa3" (UID: "e2590801-a8b2-442a-af31-b03b63138aa3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.637550 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85c1cb1a-6343-490c-98a3-f11a52290a73-config" (OuterVolumeSpecName: "config") pod "85c1cb1a-6343-490c-98a3-f11a52290a73" (UID: "85c1cb1a-6343-490c-98a3-f11a52290a73"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.640462 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85c1cb1a-6343-490c-98a3-f11a52290a73-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "85c1cb1a-6343-490c-98a3-f11a52290a73" (UID: "85c1cb1a-6343-490c-98a3-f11a52290a73"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.640492 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85c1cb1a-6343-490c-98a3-f11a52290a73-kube-api-access-49t2v" (OuterVolumeSpecName: "kube-api-access-49t2v") pod "85c1cb1a-6343-490c-98a3-f11a52290a73" (UID: "85c1cb1a-6343-490c-98a3-f11a52290a73"). InnerVolumeSpecName "kube-api-access-49t2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.645891 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2590801-a8b2-442a-af31-b03b63138aa3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e2590801-a8b2-442a-af31-b03b63138aa3" (UID: "e2590801-a8b2-442a-af31-b03b63138aa3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.646096 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2590801-a8b2-442a-af31-b03b63138aa3-kube-api-access-m2xhl" (OuterVolumeSpecName: "kube-api-access-m2xhl") pod "e2590801-a8b2-442a-af31-b03b63138aa3" (UID: "e2590801-a8b2-442a-af31-b03b63138aa3"). InnerVolumeSpecName "kube-api-access-m2xhl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.735780 4896 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/85c1cb1a-6343-490c-98a3-f11a52290a73-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.735845 4896 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85c1cb1a-6343-490c-98a3-f11a52290a73-config\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.735897 4896 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2590801-a8b2-442a-af31-b03b63138aa3-config\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.735906 4896 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e2590801-a8b2-442a-af31-b03b63138aa3-client-ca\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.735916 4896 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e2590801-a8b2-442a-af31-b03b63138aa3-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.735924 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2xhl\" (UniqueName: \"kubernetes.io/projected/e2590801-a8b2-442a-af31-b03b63138aa3-kube-api-access-m2xhl\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.735935 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49t2v\" (UniqueName: \"kubernetes.io/projected/85c1cb1a-6343-490c-98a3-f11a52290a73-kube-api-access-49t2v\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.735943 4896 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e2590801-a8b2-442a-af31-b03b63138aa3-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.735951 4896 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/85c1cb1a-6343-490c-98a3-f11a52290a73-client-ca\") on node \"crc\" DevicePath \"\"" Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.832584 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-c6db4b444-cr7c9"] Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.839835 4896 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-c6db4b444-cr7c9"] Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.844424 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c887bd576-rw68b"] Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.850115 4896 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c887bd576-rw68b"] Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.935222 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85c1cb1a-6343-490c-98a3-f11a52290a73" path="/var/lib/kubelet/pods/85c1cb1a-6343-490c-98a3-f11a52290a73/volumes" Feb 18 14:59:51 crc kubenswrapper[4896]: I0218 14:59:51.936525 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2590801-a8b2-442a-af31-b03b63138aa3" path="/var/lib/kubelet/pods/e2590801-a8b2-442a-af31-b03b63138aa3/volumes" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.061159 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6b76789c78-zbr4p"] Feb 18 14:59:52 crc kubenswrapper[4896]: E0218 14:59:52.061465 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2590801-a8b2-442a-af31-b03b63138aa3" containerName="controller-manager" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.061487 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2590801-a8b2-442a-af31-b03b63138aa3" containerName="controller-manager" Feb 18 14:59:52 crc kubenswrapper[4896]: E0218 14:59:52.061501 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85c1cb1a-6343-490c-98a3-f11a52290a73" containerName="route-controller-manager" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.061507 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="85c1cb1a-6343-490c-98a3-f11a52290a73" containerName="route-controller-manager" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.061595 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="85c1cb1a-6343-490c-98a3-f11a52290a73" containerName="route-controller-manager" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.061611 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2590801-a8b2-442a-af31-b03b63138aa3" containerName="controller-manager" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.062285 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b76789c78-zbr4p" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.063203 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cddd599db-6kmz8"] Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.063952 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cddd599db-6kmz8" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.070934 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.071600 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.072027 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.073732 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.073787 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.073839 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.073801 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.073893 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.073946 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.074020 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.074376 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.074983 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.075511 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6b76789c78-zbr4p"] Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.077141 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.078064 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cddd599db-6kmz8"] Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.142019 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9w4b\" (UniqueName: \"kubernetes.io/projected/3b1c56d7-32bf-4826-9752-2fe0e3512143-kube-api-access-m9w4b\") pod \"route-controller-manager-cddd599db-6kmz8\" (UID: \"3b1c56d7-32bf-4826-9752-2fe0e3512143\") " pod="openshift-route-controller-manager/route-controller-manager-cddd599db-6kmz8" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.142579 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3b1c56d7-32bf-4826-9752-2fe0e3512143-serving-cert\") pod \"route-controller-manager-cddd599db-6kmz8\" (UID: \"3b1c56d7-32bf-4826-9752-2fe0e3512143\") " pod="openshift-route-controller-manager/route-controller-manager-cddd599db-6kmz8" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.142732 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e1dc4728-acbf-4549-848a-50cab578d31e-client-ca\") pod \"controller-manager-6b76789c78-zbr4p\" (UID: \"e1dc4728-acbf-4549-848a-50cab578d31e\") " pod="openshift-controller-manager/controller-manager-6b76789c78-zbr4p" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.142955 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e1dc4728-acbf-4549-848a-50cab578d31e-proxy-ca-bundles\") pod \"controller-manager-6b76789c78-zbr4p\" (UID: \"e1dc4728-acbf-4549-848a-50cab578d31e\") " pod="openshift-controller-manager/controller-manager-6b76789c78-zbr4p" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.143103 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3b1c56d7-32bf-4826-9752-2fe0e3512143-client-ca\") pod \"route-controller-manager-cddd599db-6kmz8\" (UID: \"3b1c56d7-32bf-4826-9752-2fe0e3512143\") " pod="openshift-route-controller-manager/route-controller-manager-cddd599db-6kmz8" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.143300 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1dc4728-acbf-4549-848a-50cab578d31e-config\") pod \"controller-manager-6b76789c78-zbr4p\" (UID: \"e1dc4728-acbf-4549-848a-50cab578d31e\") " pod="openshift-controller-manager/controller-manager-6b76789c78-zbr4p" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.143449 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xp8sc\" (UniqueName: \"kubernetes.io/projected/e1dc4728-acbf-4549-848a-50cab578d31e-kube-api-access-xp8sc\") pod \"controller-manager-6b76789c78-zbr4p\" (UID: \"e1dc4728-acbf-4549-848a-50cab578d31e\") " pod="openshift-controller-manager/controller-manager-6b76789c78-zbr4p" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.143600 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1dc4728-acbf-4549-848a-50cab578d31e-serving-cert\") pod \"controller-manager-6b76789c78-zbr4p\" (UID: \"e1dc4728-acbf-4549-848a-50cab578d31e\") " pod="openshift-controller-manager/controller-manager-6b76789c78-zbr4p" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.143755 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b1c56d7-32bf-4826-9752-2fe0e3512143-config\") pod \"route-controller-manager-cddd599db-6kmz8\" (UID: \"3b1c56d7-32bf-4826-9752-2fe0e3512143\") " pod="openshift-route-controller-manager/route-controller-manager-cddd599db-6kmz8" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.245972 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e1dc4728-acbf-4549-848a-50cab578d31e-client-ca\") pod \"controller-manager-6b76789c78-zbr4p\" (UID: \"e1dc4728-acbf-4549-848a-50cab578d31e\") " pod="openshift-controller-manager/controller-manager-6b76789c78-zbr4p" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.246023 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e1dc4728-acbf-4549-848a-50cab578d31e-proxy-ca-bundles\") pod \"controller-manager-6b76789c78-zbr4p\" (UID: \"e1dc4728-acbf-4549-848a-50cab578d31e\") " pod="openshift-controller-manager/controller-manager-6b76789c78-zbr4p" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.246044 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3b1c56d7-32bf-4826-9752-2fe0e3512143-client-ca\") pod \"route-controller-manager-cddd599db-6kmz8\" (UID: \"3b1c56d7-32bf-4826-9752-2fe0e3512143\") " pod="openshift-route-controller-manager/route-controller-manager-cddd599db-6kmz8" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.246086 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1dc4728-acbf-4549-848a-50cab578d31e-config\") pod \"controller-manager-6b76789c78-zbr4p\" (UID: \"e1dc4728-acbf-4549-848a-50cab578d31e\") " pod="openshift-controller-manager/controller-manager-6b76789c78-zbr4p" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.246114 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xp8sc\" (UniqueName: \"kubernetes.io/projected/e1dc4728-acbf-4549-848a-50cab578d31e-kube-api-access-xp8sc\") pod \"controller-manager-6b76789c78-zbr4p\" (UID: \"e1dc4728-acbf-4549-848a-50cab578d31e\") " pod="openshift-controller-manager/controller-manager-6b76789c78-zbr4p" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.246134 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1dc4728-acbf-4549-848a-50cab578d31e-serving-cert\") pod \"controller-manager-6b76789c78-zbr4p\" (UID: \"e1dc4728-acbf-4549-848a-50cab578d31e\") " pod="openshift-controller-manager/controller-manager-6b76789c78-zbr4p" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.246150 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b1c56d7-32bf-4826-9752-2fe0e3512143-config\") pod \"route-controller-manager-cddd599db-6kmz8\" (UID: \"3b1c56d7-32bf-4826-9752-2fe0e3512143\") " pod="openshift-route-controller-manager/route-controller-manager-cddd599db-6kmz8" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.246181 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9w4b\" (UniqueName: \"kubernetes.io/projected/3b1c56d7-32bf-4826-9752-2fe0e3512143-kube-api-access-m9w4b\") pod \"route-controller-manager-cddd599db-6kmz8\" (UID: \"3b1c56d7-32bf-4826-9752-2fe0e3512143\") " pod="openshift-route-controller-manager/route-controller-manager-cddd599db-6kmz8" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.246195 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3b1c56d7-32bf-4826-9752-2fe0e3512143-serving-cert\") pod \"route-controller-manager-cddd599db-6kmz8\" (UID: \"3b1c56d7-32bf-4826-9752-2fe0e3512143\") " pod="openshift-route-controller-manager/route-controller-manager-cddd599db-6kmz8" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.249316 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1dc4728-acbf-4549-848a-50cab578d31e-config\") pod \"controller-manager-6b76789c78-zbr4p\" (UID: \"e1dc4728-acbf-4549-848a-50cab578d31e\") " pod="openshift-controller-manager/controller-manager-6b76789c78-zbr4p" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.249888 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e1dc4728-acbf-4549-848a-50cab578d31e-proxy-ca-bundles\") pod \"controller-manager-6b76789c78-zbr4p\" (UID: \"e1dc4728-acbf-4549-848a-50cab578d31e\") " pod="openshift-controller-manager/controller-manager-6b76789c78-zbr4p" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.249952 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3b1c56d7-32bf-4826-9752-2fe0e3512143-serving-cert\") pod \"route-controller-manager-cddd599db-6kmz8\" (UID: \"3b1c56d7-32bf-4826-9752-2fe0e3512143\") " pod="openshift-route-controller-manager/route-controller-manager-cddd599db-6kmz8" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.250551 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3b1c56d7-32bf-4826-9752-2fe0e3512143-client-ca\") pod \"route-controller-manager-cddd599db-6kmz8\" (UID: \"3b1c56d7-32bf-4826-9752-2fe0e3512143\") " pod="openshift-route-controller-manager/route-controller-manager-cddd599db-6kmz8" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.250785 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b1c56d7-32bf-4826-9752-2fe0e3512143-config\") pod \"route-controller-manager-cddd599db-6kmz8\" (UID: \"3b1c56d7-32bf-4826-9752-2fe0e3512143\") " pod="openshift-route-controller-manager/route-controller-manager-cddd599db-6kmz8" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.251227 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e1dc4728-acbf-4549-848a-50cab578d31e-client-ca\") pod \"controller-manager-6b76789c78-zbr4p\" (UID: \"e1dc4728-acbf-4549-848a-50cab578d31e\") " pod="openshift-controller-manager/controller-manager-6b76789c78-zbr4p" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.252352 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1dc4728-acbf-4549-848a-50cab578d31e-serving-cert\") pod \"controller-manager-6b76789c78-zbr4p\" (UID: \"e1dc4728-acbf-4549-848a-50cab578d31e\") " pod="openshift-controller-manager/controller-manager-6b76789c78-zbr4p" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.269388 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xp8sc\" (UniqueName: \"kubernetes.io/projected/e1dc4728-acbf-4549-848a-50cab578d31e-kube-api-access-xp8sc\") pod \"controller-manager-6b76789c78-zbr4p\" (UID: \"e1dc4728-acbf-4549-848a-50cab578d31e\") " pod="openshift-controller-manager/controller-manager-6b76789c78-zbr4p" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.273864 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9w4b\" (UniqueName: \"kubernetes.io/projected/3b1c56d7-32bf-4826-9752-2fe0e3512143-kube-api-access-m9w4b\") pod \"route-controller-manager-cddd599db-6kmz8\" (UID: \"3b1c56d7-32bf-4826-9752-2fe0e3512143\") " pod="openshift-route-controller-manager/route-controller-manager-cddd599db-6kmz8" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.391508 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b76789c78-zbr4p" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.414845 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cddd599db-6kmz8" Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.585635 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6b76789c78-zbr4p"] Feb 18 14:59:52 crc kubenswrapper[4896]: I0218 14:59:52.631039 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cddd599db-6kmz8"] Feb 18 14:59:52 crc kubenswrapper[4896]: W0218 14:59:52.643085 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b1c56d7_32bf_4826_9752_2fe0e3512143.slice/crio-159a9faf9ec30bd16aac877f99bff7b671eee9cbc5fb34d8babc0b5d35b3144a WatchSource:0}: Error finding container 159a9faf9ec30bd16aac877f99bff7b671eee9cbc5fb34d8babc0b5d35b3144a: Status 404 returned error can't find the container with id 159a9faf9ec30bd16aac877f99bff7b671eee9cbc5fb34d8babc0b5d35b3144a Feb 18 14:59:53 crc kubenswrapper[4896]: I0218 14:59:53.512450 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cddd599db-6kmz8" event={"ID":"3b1c56d7-32bf-4826-9752-2fe0e3512143","Type":"ContainerStarted","Data":"d668ecef1be10ca5c3c33d6e7f26eb8f528580361d115a2303598f82f870deaa"} Feb 18 14:59:53 crc kubenswrapper[4896]: I0218 14:59:53.512777 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-cddd599db-6kmz8" Feb 18 14:59:53 crc kubenswrapper[4896]: I0218 14:59:53.512789 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cddd599db-6kmz8" event={"ID":"3b1c56d7-32bf-4826-9752-2fe0e3512143","Type":"ContainerStarted","Data":"159a9faf9ec30bd16aac877f99bff7b671eee9cbc5fb34d8babc0b5d35b3144a"} Feb 18 14:59:53 crc kubenswrapper[4896]: I0218 14:59:53.514262 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b76789c78-zbr4p" event={"ID":"e1dc4728-acbf-4549-848a-50cab578d31e","Type":"ContainerStarted","Data":"f27d2f4120dda872d47adbbadb451c3729cc04a9b805f01beeefad73209f7382"} Feb 18 14:59:53 crc kubenswrapper[4896]: I0218 14:59:53.514303 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b76789c78-zbr4p" event={"ID":"e1dc4728-acbf-4549-848a-50cab578d31e","Type":"ContainerStarted","Data":"b6a36e401b49f96e810b9ab3da1c0f70b13d7b0afed12e5c29b55bcf62ed9d5d"} Feb 18 14:59:53 crc kubenswrapper[4896]: I0218 14:59:53.514428 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6b76789c78-zbr4p" Feb 18 14:59:53 crc kubenswrapper[4896]: I0218 14:59:53.517891 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6b76789c78-zbr4p" Feb 18 14:59:53 crc kubenswrapper[4896]: I0218 14:59:53.517943 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-cddd599db-6kmz8" Feb 18 14:59:53 crc kubenswrapper[4896]: I0218 14:59:53.534125 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-cddd599db-6kmz8" podStartSLOduration=3.534099378 podStartE2EDuration="3.534099378s" podCreationTimestamp="2026-02-18 14:59:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:59:53.533673686 +0000 UTC m=+292.463007909" watchObservedRunningTime="2026-02-18 14:59:53.534099378 +0000 UTC m=+292.463433601" Feb 18 14:59:53 crc kubenswrapper[4896]: I0218 14:59:53.548078 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6b76789c78-zbr4p" podStartSLOduration=3.548060063 podStartE2EDuration="3.548060063s" podCreationTimestamp="2026-02-18 14:59:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 14:59:53.546193251 +0000 UTC m=+292.475527464" watchObservedRunningTime="2026-02-18 14:59:53.548060063 +0000 UTC m=+292.477394296" Feb 18 15:00:00 crc kubenswrapper[4896]: I0218 15:00:00.159912 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523780-5njls"] Feb 18 15:00:00 crc kubenswrapper[4896]: I0218 15:00:00.160956 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523780-5njls" Feb 18 15:00:00 crc kubenswrapper[4896]: I0218 15:00:00.163796 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 18 15:00:00 crc kubenswrapper[4896]: I0218 15:00:00.164020 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 18 15:00:00 crc kubenswrapper[4896]: I0218 15:00:00.170967 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523780-5njls"] Feb 18 15:00:00 crc kubenswrapper[4896]: I0218 15:00:00.235725 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0d666ffc-5489-4386-8b04-43d4e324eaad-config-volume\") pod \"collect-profiles-29523780-5njls\" (UID: \"0d666ffc-5489-4386-8b04-43d4e324eaad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523780-5njls" Feb 18 15:00:00 crc kubenswrapper[4896]: I0218 15:00:00.235850 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xj86\" (UniqueName: \"kubernetes.io/projected/0d666ffc-5489-4386-8b04-43d4e324eaad-kube-api-access-5xj86\") pod \"collect-profiles-29523780-5njls\" (UID: \"0d666ffc-5489-4386-8b04-43d4e324eaad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523780-5njls" Feb 18 15:00:00 crc kubenswrapper[4896]: I0218 15:00:00.235929 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0d666ffc-5489-4386-8b04-43d4e324eaad-secret-volume\") pod \"collect-profiles-29523780-5njls\" (UID: \"0d666ffc-5489-4386-8b04-43d4e324eaad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523780-5njls" Feb 18 15:00:00 crc kubenswrapper[4896]: I0218 15:00:00.336525 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xj86\" (UniqueName: \"kubernetes.io/projected/0d666ffc-5489-4386-8b04-43d4e324eaad-kube-api-access-5xj86\") pod \"collect-profiles-29523780-5njls\" (UID: \"0d666ffc-5489-4386-8b04-43d4e324eaad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523780-5njls" Feb 18 15:00:00 crc kubenswrapper[4896]: I0218 15:00:00.336602 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0d666ffc-5489-4386-8b04-43d4e324eaad-secret-volume\") pod \"collect-profiles-29523780-5njls\" (UID: \"0d666ffc-5489-4386-8b04-43d4e324eaad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523780-5njls" Feb 18 15:00:00 crc kubenswrapper[4896]: I0218 15:00:00.336631 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0d666ffc-5489-4386-8b04-43d4e324eaad-config-volume\") pod \"collect-profiles-29523780-5njls\" (UID: \"0d666ffc-5489-4386-8b04-43d4e324eaad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523780-5njls" Feb 18 15:00:00 crc kubenswrapper[4896]: I0218 15:00:00.337508 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0d666ffc-5489-4386-8b04-43d4e324eaad-config-volume\") pod \"collect-profiles-29523780-5njls\" (UID: \"0d666ffc-5489-4386-8b04-43d4e324eaad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523780-5njls" Feb 18 15:00:00 crc kubenswrapper[4896]: I0218 15:00:00.342480 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0d666ffc-5489-4386-8b04-43d4e324eaad-secret-volume\") pod \"collect-profiles-29523780-5njls\" (UID: \"0d666ffc-5489-4386-8b04-43d4e324eaad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523780-5njls" Feb 18 15:00:00 crc kubenswrapper[4896]: I0218 15:00:00.354066 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xj86\" (UniqueName: \"kubernetes.io/projected/0d666ffc-5489-4386-8b04-43d4e324eaad-kube-api-access-5xj86\") pod \"collect-profiles-29523780-5njls\" (UID: \"0d666ffc-5489-4386-8b04-43d4e324eaad\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523780-5njls" Feb 18 15:00:00 crc kubenswrapper[4896]: I0218 15:00:00.483701 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523780-5njls" Feb 18 15:00:00 crc kubenswrapper[4896]: I0218 15:00:00.550586 4896 generic.go:334] "Generic (PLEG): container finished" podID="0f95dce4-6d68-44ac-9dd3-215664bb1d75" containerID="95e1564d67b45208daa6c2d917ef9ddc045a644295fdcc1557bdf6c9fdbea897" exitCode=0 Feb 18 15:00:00 crc kubenswrapper[4896]: I0218 15:00:00.550675 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xrhcj" event={"ID":"0f95dce4-6d68-44ac-9dd3-215664bb1d75","Type":"ContainerDied","Data":"95e1564d67b45208daa6c2d917ef9ddc045a644295fdcc1557bdf6c9fdbea897"} Feb 18 15:00:00 crc kubenswrapper[4896]: I0218 15:00:00.551601 4896 scope.go:117] "RemoveContainer" containerID="95e1564d67b45208daa6c2d917ef9ddc045a644295fdcc1557bdf6c9fdbea897" Feb 18 15:00:00 crc kubenswrapper[4896]: I0218 15:00:00.904542 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523780-5njls"] Feb 18 15:00:01 crc kubenswrapper[4896]: I0218 15:00:01.529545 4896 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Feb 18 15:00:01 crc kubenswrapper[4896]: I0218 15:00:01.557526 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xrhcj" event={"ID":"0f95dce4-6d68-44ac-9dd3-215664bb1d75","Type":"ContainerStarted","Data":"3ae5e1d63464b4ac3259ba7f756992398a78f3fd9bf860514b62bb5252ea5fb1"} Feb 18 15:00:01 crc kubenswrapper[4896]: I0218 15:00:01.557819 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-xrhcj" Feb 18 15:00:01 crc kubenswrapper[4896]: I0218 15:00:01.560155 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-xrhcj" Feb 18 15:00:01 crc kubenswrapper[4896]: I0218 15:00:01.560531 4896 generic.go:334] "Generic (PLEG): container finished" podID="0d666ffc-5489-4386-8b04-43d4e324eaad" containerID="fce7481ade1dacb6329c8087564357078451dbb7e4b6998f4d196886e97c1791" exitCode=0 Feb 18 15:00:01 crc kubenswrapper[4896]: I0218 15:00:01.560573 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523780-5njls" event={"ID":"0d666ffc-5489-4386-8b04-43d4e324eaad","Type":"ContainerDied","Data":"fce7481ade1dacb6329c8087564357078451dbb7e4b6998f4d196886e97c1791"} Feb 18 15:00:01 crc kubenswrapper[4896]: I0218 15:00:01.560633 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523780-5njls" event={"ID":"0d666ffc-5489-4386-8b04-43d4e324eaad","Type":"ContainerStarted","Data":"fc1d60038a21d979b024a8adeeea99551ad2c264efcbcba1e6247094ae6a3af1"} Feb 18 15:00:02 crc kubenswrapper[4896]: I0218 15:00:02.878939 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523780-5njls" Feb 18 15:00:03 crc kubenswrapper[4896]: I0218 15:00:03.069499 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0d666ffc-5489-4386-8b04-43d4e324eaad-secret-volume\") pod \"0d666ffc-5489-4386-8b04-43d4e324eaad\" (UID: \"0d666ffc-5489-4386-8b04-43d4e324eaad\") " Feb 18 15:00:03 crc kubenswrapper[4896]: I0218 15:00:03.069565 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xj86\" (UniqueName: \"kubernetes.io/projected/0d666ffc-5489-4386-8b04-43d4e324eaad-kube-api-access-5xj86\") pod \"0d666ffc-5489-4386-8b04-43d4e324eaad\" (UID: \"0d666ffc-5489-4386-8b04-43d4e324eaad\") " Feb 18 15:00:03 crc kubenswrapper[4896]: I0218 15:00:03.069642 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0d666ffc-5489-4386-8b04-43d4e324eaad-config-volume\") pod \"0d666ffc-5489-4386-8b04-43d4e324eaad\" (UID: \"0d666ffc-5489-4386-8b04-43d4e324eaad\") " Feb 18 15:00:03 crc kubenswrapper[4896]: I0218 15:00:03.070959 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d666ffc-5489-4386-8b04-43d4e324eaad-config-volume" (OuterVolumeSpecName: "config-volume") pod "0d666ffc-5489-4386-8b04-43d4e324eaad" (UID: "0d666ffc-5489-4386-8b04-43d4e324eaad"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:00:03 crc kubenswrapper[4896]: I0218 15:00:03.075704 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d666ffc-5489-4386-8b04-43d4e324eaad-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0d666ffc-5489-4386-8b04-43d4e324eaad" (UID: "0d666ffc-5489-4386-8b04-43d4e324eaad"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:00:03 crc kubenswrapper[4896]: I0218 15:00:03.076402 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d666ffc-5489-4386-8b04-43d4e324eaad-kube-api-access-5xj86" (OuterVolumeSpecName: "kube-api-access-5xj86") pod "0d666ffc-5489-4386-8b04-43d4e324eaad" (UID: "0d666ffc-5489-4386-8b04-43d4e324eaad"). InnerVolumeSpecName "kube-api-access-5xj86". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:00:03 crc kubenswrapper[4896]: I0218 15:00:03.170699 4896 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0d666ffc-5489-4386-8b04-43d4e324eaad-config-volume\") on node \"crc\" DevicePath \"\"" Feb 18 15:00:03 crc kubenswrapper[4896]: I0218 15:00:03.170736 4896 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0d666ffc-5489-4386-8b04-43d4e324eaad-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 18 15:00:03 crc kubenswrapper[4896]: I0218 15:00:03.170746 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xj86\" (UniqueName: \"kubernetes.io/projected/0d666ffc-5489-4386-8b04-43d4e324eaad-kube-api-access-5xj86\") on node \"crc\" DevicePath \"\"" Feb 18 15:00:03 crc kubenswrapper[4896]: I0218 15:00:03.570743 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523780-5njls" event={"ID":"0d666ffc-5489-4386-8b04-43d4e324eaad","Type":"ContainerDied","Data":"fc1d60038a21d979b024a8adeeea99551ad2c264efcbcba1e6247094ae6a3af1"} Feb 18 15:00:03 crc kubenswrapper[4896]: I0218 15:00:03.571033 4896 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc1d60038a21d979b024a8adeeea99551ad2c264efcbcba1e6247094ae6a3af1" Feb 18 15:00:03 crc kubenswrapper[4896]: I0218 15:00:03.570764 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523780-5njls" Feb 18 15:00:10 crc kubenswrapper[4896]: I0218 15:00:10.501859 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6b76789c78-zbr4p"] Feb 18 15:00:10 crc kubenswrapper[4896]: I0218 15:00:10.502659 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6b76789c78-zbr4p" podUID="e1dc4728-acbf-4549-848a-50cab578d31e" containerName="controller-manager" containerID="cri-o://f27d2f4120dda872d47adbbadb451c3729cc04a9b805f01beeefad73209f7382" gracePeriod=30 Feb 18 15:00:10 crc kubenswrapper[4896]: I0218 15:00:10.517069 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cddd599db-6kmz8"] Feb 18 15:00:10 crc kubenswrapper[4896]: I0218 15:00:10.517312 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-cddd599db-6kmz8" podUID="3b1c56d7-32bf-4826-9752-2fe0e3512143" containerName="route-controller-manager" containerID="cri-o://d668ecef1be10ca5c3c33d6e7f26eb8f528580361d115a2303598f82f870deaa" gracePeriod=30 Feb 18 15:00:10 crc kubenswrapper[4896]: I0218 15:00:10.995032 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cddd599db-6kmz8" Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.108950 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b76789c78-zbr4p" Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.160083 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3b1c56d7-32bf-4826-9752-2fe0e3512143-serving-cert\") pod \"3b1c56d7-32bf-4826-9752-2fe0e3512143\" (UID: \"3b1c56d7-32bf-4826-9752-2fe0e3512143\") " Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.161007 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b1c56d7-32bf-4826-9752-2fe0e3512143-config" (OuterVolumeSpecName: "config") pod "3b1c56d7-32bf-4826-9752-2fe0e3512143" (UID: "3b1c56d7-32bf-4826-9752-2fe0e3512143"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.161067 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b1c56d7-32bf-4826-9752-2fe0e3512143-config\") pod \"3b1c56d7-32bf-4826-9752-2fe0e3512143\" (UID: \"3b1c56d7-32bf-4826-9752-2fe0e3512143\") " Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.161123 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9w4b\" (UniqueName: \"kubernetes.io/projected/3b1c56d7-32bf-4826-9752-2fe0e3512143-kube-api-access-m9w4b\") pod \"3b1c56d7-32bf-4826-9752-2fe0e3512143\" (UID: \"3b1c56d7-32bf-4826-9752-2fe0e3512143\") " Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.161598 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b1c56d7-32bf-4826-9752-2fe0e3512143-client-ca" (OuterVolumeSpecName: "client-ca") pod "3b1c56d7-32bf-4826-9752-2fe0e3512143" (UID: "3b1c56d7-32bf-4826-9752-2fe0e3512143"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.161165 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3b1c56d7-32bf-4826-9752-2fe0e3512143-client-ca\") pod \"3b1c56d7-32bf-4826-9752-2fe0e3512143\" (UID: \"3b1c56d7-32bf-4826-9752-2fe0e3512143\") " Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.161818 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xp8sc\" (UniqueName: \"kubernetes.io/projected/e1dc4728-acbf-4549-848a-50cab578d31e-kube-api-access-xp8sc\") pod \"e1dc4728-acbf-4549-848a-50cab578d31e\" (UID: \"e1dc4728-acbf-4549-848a-50cab578d31e\") " Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.161997 4896 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b1c56d7-32bf-4826-9752-2fe0e3512143-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.162014 4896 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3b1c56d7-32bf-4826-9752-2fe0e3512143-client-ca\") on node \"crc\" DevicePath \"\"" Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.164822 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1dc4728-acbf-4549-848a-50cab578d31e-kube-api-access-xp8sc" (OuterVolumeSpecName: "kube-api-access-xp8sc") pod "e1dc4728-acbf-4549-848a-50cab578d31e" (UID: "e1dc4728-acbf-4549-848a-50cab578d31e"). InnerVolumeSpecName "kube-api-access-xp8sc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.164843 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b1c56d7-32bf-4826-9752-2fe0e3512143-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "3b1c56d7-32bf-4826-9752-2fe0e3512143" (UID: "3b1c56d7-32bf-4826-9752-2fe0e3512143"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.166412 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b1c56d7-32bf-4826-9752-2fe0e3512143-kube-api-access-m9w4b" (OuterVolumeSpecName: "kube-api-access-m9w4b") pod "3b1c56d7-32bf-4826-9752-2fe0e3512143" (UID: "3b1c56d7-32bf-4826-9752-2fe0e3512143"). InnerVolumeSpecName "kube-api-access-m9w4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.262570 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e1dc4728-acbf-4549-848a-50cab578d31e-proxy-ca-bundles\") pod \"e1dc4728-acbf-4549-848a-50cab578d31e\" (UID: \"e1dc4728-acbf-4549-848a-50cab578d31e\") " Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.262624 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1dc4728-acbf-4549-848a-50cab578d31e-serving-cert\") pod \"e1dc4728-acbf-4549-848a-50cab578d31e\" (UID: \"e1dc4728-acbf-4549-848a-50cab578d31e\") " Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.262664 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1dc4728-acbf-4549-848a-50cab578d31e-config\") pod \"e1dc4728-acbf-4549-848a-50cab578d31e\" (UID: \"e1dc4728-acbf-4549-848a-50cab578d31e\") " Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.262710 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e1dc4728-acbf-4549-848a-50cab578d31e-client-ca\") pod \"e1dc4728-acbf-4549-848a-50cab578d31e\" (UID: \"e1dc4728-acbf-4549-848a-50cab578d31e\") " Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.262892 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9w4b\" (UniqueName: \"kubernetes.io/projected/3b1c56d7-32bf-4826-9752-2fe0e3512143-kube-api-access-m9w4b\") on node \"crc\" DevicePath \"\"" Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.262904 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xp8sc\" (UniqueName: \"kubernetes.io/projected/e1dc4728-acbf-4549-848a-50cab578d31e-kube-api-access-xp8sc\") on node \"crc\" DevicePath \"\"" Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.262913 4896 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3b1c56d7-32bf-4826-9752-2fe0e3512143-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.263289 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1dc4728-acbf-4549-848a-50cab578d31e-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "e1dc4728-acbf-4549-848a-50cab578d31e" (UID: "e1dc4728-acbf-4549-848a-50cab578d31e"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.263308 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1dc4728-acbf-4549-848a-50cab578d31e-client-ca" (OuterVolumeSpecName: "client-ca") pod "e1dc4728-acbf-4549-848a-50cab578d31e" (UID: "e1dc4728-acbf-4549-848a-50cab578d31e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.263324 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1dc4728-acbf-4549-848a-50cab578d31e-config" (OuterVolumeSpecName: "config") pod "e1dc4728-acbf-4549-848a-50cab578d31e" (UID: "e1dc4728-acbf-4549-848a-50cab578d31e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.265580 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1dc4728-acbf-4549-848a-50cab578d31e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e1dc4728-acbf-4549-848a-50cab578d31e" (UID: "e1dc4728-acbf-4549-848a-50cab578d31e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.363262 4896 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e1dc4728-acbf-4549-848a-50cab578d31e-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.363293 4896 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1dc4728-acbf-4549-848a-50cab578d31e-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.363305 4896 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1dc4728-acbf-4549-848a-50cab578d31e-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.363314 4896 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e1dc4728-acbf-4549-848a-50cab578d31e-client-ca\") on node \"crc\" DevicePath \"\"" Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.609126 4896 generic.go:334] "Generic (PLEG): container finished" podID="3b1c56d7-32bf-4826-9752-2fe0e3512143" containerID="d668ecef1be10ca5c3c33d6e7f26eb8f528580361d115a2303598f82f870deaa" exitCode=0 Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.609199 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cddd599db-6kmz8" event={"ID":"3b1c56d7-32bf-4826-9752-2fe0e3512143","Type":"ContainerDied","Data":"d668ecef1be10ca5c3c33d6e7f26eb8f528580361d115a2303598f82f870deaa"} Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.609274 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-cddd599db-6kmz8" event={"ID":"3b1c56d7-32bf-4826-9752-2fe0e3512143","Type":"ContainerDied","Data":"159a9faf9ec30bd16aac877f99bff7b671eee9cbc5fb34d8babc0b5d35b3144a"} Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.609301 4896 scope.go:117] "RemoveContainer" containerID="d668ecef1be10ca5c3c33d6e7f26eb8f528580361d115a2303598f82f870deaa" Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.609469 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-cddd599db-6kmz8" Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.610522 4896 generic.go:334] "Generic (PLEG): container finished" podID="e1dc4728-acbf-4549-848a-50cab578d31e" containerID="f27d2f4120dda872d47adbbadb451c3729cc04a9b805f01beeefad73209f7382" exitCode=0 Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.610549 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b76789c78-zbr4p" event={"ID":"e1dc4728-acbf-4549-848a-50cab578d31e","Type":"ContainerDied","Data":"f27d2f4120dda872d47adbbadb451c3729cc04a9b805f01beeefad73209f7382"} Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.610567 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6b76789c78-zbr4p" event={"ID":"e1dc4728-acbf-4549-848a-50cab578d31e","Type":"ContainerDied","Data":"b6a36e401b49f96e810b9ab3da1c0f70b13d7b0afed12e5c29b55bcf62ed9d5d"} Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.610656 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6b76789c78-zbr4p" Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.620874 4896 scope.go:117] "RemoveContainer" containerID="d668ecef1be10ca5c3c33d6e7f26eb8f528580361d115a2303598f82f870deaa" Feb 18 15:00:11 crc kubenswrapper[4896]: E0218 15:00:11.621955 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d668ecef1be10ca5c3c33d6e7f26eb8f528580361d115a2303598f82f870deaa\": container with ID starting with d668ecef1be10ca5c3c33d6e7f26eb8f528580361d115a2303598f82f870deaa not found: ID does not exist" containerID="d668ecef1be10ca5c3c33d6e7f26eb8f528580361d115a2303598f82f870deaa" Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.623356 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d668ecef1be10ca5c3c33d6e7f26eb8f528580361d115a2303598f82f870deaa"} err="failed to get container status \"d668ecef1be10ca5c3c33d6e7f26eb8f528580361d115a2303598f82f870deaa\": rpc error: code = NotFound desc = could not find container \"d668ecef1be10ca5c3c33d6e7f26eb8f528580361d115a2303598f82f870deaa\": container with ID starting with d668ecef1be10ca5c3c33d6e7f26eb8f528580361d115a2303598f82f870deaa not found: ID does not exist" Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.623417 4896 scope.go:117] "RemoveContainer" containerID="f27d2f4120dda872d47adbbadb451c3729cc04a9b805f01beeefad73209f7382" Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.637221 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6b76789c78-zbr4p"] Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.640584 4896 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6b76789c78-zbr4p"] Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.641391 4896 scope.go:117] "RemoveContainer" containerID="f27d2f4120dda872d47adbbadb451c3729cc04a9b805f01beeefad73209f7382" Feb 18 15:00:11 crc kubenswrapper[4896]: E0218 15:00:11.641845 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f27d2f4120dda872d47adbbadb451c3729cc04a9b805f01beeefad73209f7382\": container with ID starting with f27d2f4120dda872d47adbbadb451c3729cc04a9b805f01beeefad73209f7382 not found: ID does not exist" containerID="f27d2f4120dda872d47adbbadb451c3729cc04a9b805f01beeefad73209f7382" Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.641887 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f27d2f4120dda872d47adbbadb451c3729cc04a9b805f01beeefad73209f7382"} err="failed to get container status \"f27d2f4120dda872d47adbbadb451c3729cc04a9b805f01beeefad73209f7382\": rpc error: code = NotFound desc = could not find container \"f27d2f4120dda872d47adbbadb451c3729cc04a9b805f01beeefad73209f7382\": container with ID starting with f27d2f4120dda872d47adbbadb451c3729cc04a9b805f01beeefad73209f7382 not found: ID does not exist" Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.646170 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cddd599db-6kmz8"] Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.650296 4896 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-cddd599db-6kmz8"] Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.934144 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b1c56d7-32bf-4826-9752-2fe0e3512143" path="/var/lib/kubelet/pods/3b1c56d7-32bf-4826-9752-2fe0e3512143/volumes" Feb 18 15:00:11 crc kubenswrapper[4896]: I0218 15:00:11.934849 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1dc4728-acbf-4549-848a-50cab578d31e" path="/var/lib/kubelet/pods/e1dc4728-acbf-4549-848a-50cab578d31e/volumes" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.069203 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5f69f9958d-lwmg4"] Feb 18 15:00:12 crc kubenswrapper[4896]: E0218 15:00:12.069499 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d666ffc-5489-4386-8b04-43d4e324eaad" containerName="collect-profiles" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.069517 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d666ffc-5489-4386-8b04-43d4e324eaad" containerName="collect-profiles" Feb 18 15:00:12 crc kubenswrapper[4896]: E0218 15:00:12.069533 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b1c56d7-32bf-4826-9752-2fe0e3512143" containerName="route-controller-manager" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.069540 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b1c56d7-32bf-4826-9752-2fe0e3512143" containerName="route-controller-manager" Feb 18 15:00:12 crc kubenswrapper[4896]: E0218 15:00:12.069551 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1dc4728-acbf-4549-848a-50cab578d31e" containerName="controller-manager" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.069558 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1dc4728-acbf-4549-848a-50cab578d31e" containerName="controller-manager" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.069644 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d666ffc-5489-4386-8b04-43d4e324eaad" containerName="collect-profiles" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.069655 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1dc4728-acbf-4549-848a-50cab578d31e" containerName="controller-manager" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.069668 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b1c56d7-32bf-4826-9752-2fe0e3512143" containerName="route-controller-manager" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.070036 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f69f9958d-lwmg4" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.072327 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e8f7e30b-d57f-4849-8750-bc501da6b0f6-serving-cert\") pod \"controller-manager-5f69f9958d-lwmg4\" (UID: \"e8f7e30b-d57f-4849-8750-bc501da6b0f6\") " pod="openshift-controller-manager/controller-manager-5f69f9958d-lwmg4" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.072397 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8ffn\" (UniqueName: \"kubernetes.io/projected/e8f7e30b-d57f-4849-8750-bc501da6b0f6-kube-api-access-k8ffn\") pod \"controller-manager-5f69f9958d-lwmg4\" (UID: \"e8f7e30b-d57f-4849-8750-bc501da6b0f6\") " pod="openshift-controller-manager/controller-manager-5f69f9958d-lwmg4" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.072433 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8f7e30b-d57f-4849-8750-bc501da6b0f6-config\") pod \"controller-manager-5f69f9958d-lwmg4\" (UID: \"e8f7e30b-d57f-4849-8750-bc501da6b0f6\") " pod="openshift-controller-manager/controller-manager-5f69f9958d-lwmg4" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.072452 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e8f7e30b-d57f-4849-8750-bc501da6b0f6-proxy-ca-bundles\") pod \"controller-manager-5f69f9958d-lwmg4\" (UID: \"e8f7e30b-d57f-4849-8750-bc501da6b0f6\") " pod="openshift-controller-manager/controller-manager-5f69f9958d-lwmg4" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.072476 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e8f7e30b-d57f-4849-8750-bc501da6b0f6-client-ca\") pod \"controller-manager-5f69f9958d-lwmg4\" (UID: \"e8f7e30b-d57f-4849-8750-bc501da6b0f6\") " pod="openshift-controller-manager/controller-manager-5f69f9958d-lwmg4" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.072604 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.072982 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-576968d44d-lgnz7"] Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.073190 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.073258 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.073305 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.073581 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.073867 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-576968d44d-lgnz7" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.075702 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.075869 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.076008 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.076194 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.076336 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.076463 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.079367 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.081795 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.082481 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-576968d44d-lgnz7"] Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.092269 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5f69f9958d-lwmg4"] Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.173664 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8f7e30b-d57f-4849-8750-bc501da6b0f6-config\") pod \"controller-manager-5f69f9958d-lwmg4\" (UID: \"e8f7e30b-d57f-4849-8750-bc501da6b0f6\") " pod="openshift-controller-manager/controller-manager-5f69f9958d-lwmg4" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.173735 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e8f7e30b-d57f-4849-8750-bc501da6b0f6-proxy-ca-bundles\") pod \"controller-manager-5f69f9958d-lwmg4\" (UID: \"e8f7e30b-d57f-4849-8750-bc501da6b0f6\") " pod="openshift-controller-manager/controller-manager-5f69f9958d-lwmg4" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.173768 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e8f7e30b-d57f-4849-8750-bc501da6b0f6-client-ca\") pod \"controller-manager-5f69f9958d-lwmg4\" (UID: \"e8f7e30b-d57f-4849-8750-bc501da6b0f6\") " pod="openshift-controller-manager/controller-manager-5f69f9958d-lwmg4" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.173831 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4cds\" (UniqueName: \"kubernetes.io/projected/6de61e80-75b6-402a-8166-65555cd3a0f7-kube-api-access-c4cds\") pod \"route-controller-manager-576968d44d-lgnz7\" (UID: \"6de61e80-75b6-402a-8166-65555cd3a0f7\") " pod="openshift-route-controller-manager/route-controller-manager-576968d44d-lgnz7" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.173863 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e8f7e30b-d57f-4849-8750-bc501da6b0f6-serving-cert\") pod \"controller-manager-5f69f9958d-lwmg4\" (UID: \"e8f7e30b-d57f-4849-8750-bc501da6b0f6\") " pod="openshift-controller-manager/controller-manager-5f69f9958d-lwmg4" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.173905 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6de61e80-75b6-402a-8166-65555cd3a0f7-client-ca\") pod \"route-controller-manager-576968d44d-lgnz7\" (UID: \"6de61e80-75b6-402a-8166-65555cd3a0f7\") " pod="openshift-route-controller-manager/route-controller-manager-576968d44d-lgnz7" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.173927 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8ffn\" (UniqueName: \"kubernetes.io/projected/e8f7e30b-d57f-4849-8750-bc501da6b0f6-kube-api-access-k8ffn\") pod \"controller-manager-5f69f9958d-lwmg4\" (UID: \"e8f7e30b-d57f-4849-8750-bc501da6b0f6\") " pod="openshift-controller-manager/controller-manager-5f69f9958d-lwmg4" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.173944 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6de61e80-75b6-402a-8166-65555cd3a0f7-config\") pod \"route-controller-manager-576968d44d-lgnz7\" (UID: \"6de61e80-75b6-402a-8166-65555cd3a0f7\") " pod="openshift-route-controller-manager/route-controller-manager-576968d44d-lgnz7" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.173982 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6de61e80-75b6-402a-8166-65555cd3a0f7-serving-cert\") pod \"route-controller-manager-576968d44d-lgnz7\" (UID: \"6de61e80-75b6-402a-8166-65555cd3a0f7\") " pod="openshift-route-controller-manager/route-controller-manager-576968d44d-lgnz7" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.175122 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e8f7e30b-d57f-4849-8750-bc501da6b0f6-proxy-ca-bundles\") pod \"controller-manager-5f69f9958d-lwmg4\" (UID: \"e8f7e30b-d57f-4849-8750-bc501da6b0f6\") " pod="openshift-controller-manager/controller-manager-5f69f9958d-lwmg4" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.175118 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e8f7e30b-d57f-4849-8750-bc501da6b0f6-config\") pod \"controller-manager-5f69f9958d-lwmg4\" (UID: \"e8f7e30b-d57f-4849-8750-bc501da6b0f6\") " pod="openshift-controller-manager/controller-manager-5f69f9958d-lwmg4" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.175473 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e8f7e30b-d57f-4849-8750-bc501da6b0f6-client-ca\") pod \"controller-manager-5f69f9958d-lwmg4\" (UID: \"e8f7e30b-d57f-4849-8750-bc501da6b0f6\") " pod="openshift-controller-manager/controller-manager-5f69f9958d-lwmg4" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.186780 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e8f7e30b-d57f-4849-8750-bc501da6b0f6-serving-cert\") pod \"controller-manager-5f69f9958d-lwmg4\" (UID: \"e8f7e30b-d57f-4849-8750-bc501da6b0f6\") " pod="openshift-controller-manager/controller-manager-5f69f9958d-lwmg4" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.191513 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8ffn\" (UniqueName: \"kubernetes.io/projected/e8f7e30b-d57f-4849-8750-bc501da6b0f6-kube-api-access-k8ffn\") pod \"controller-manager-5f69f9958d-lwmg4\" (UID: \"e8f7e30b-d57f-4849-8750-bc501da6b0f6\") " pod="openshift-controller-manager/controller-manager-5f69f9958d-lwmg4" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.274907 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4cds\" (UniqueName: \"kubernetes.io/projected/6de61e80-75b6-402a-8166-65555cd3a0f7-kube-api-access-c4cds\") pod \"route-controller-manager-576968d44d-lgnz7\" (UID: \"6de61e80-75b6-402a-8166-65555cd3a0f7\") " pod="openshift-route-controller-manager/route-controller-manager-576968d44d-lgnz7" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.275025 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6de61e80-75b6-402a-8166-65555cd3a0f7-client-ca\") pod \"route-controller-manager-576968d44d-lgnz7\" (UID: \"6de61e80-75b6-402a-8166-65555cd3a0f7\") " pod="openshift-route-controller-manager/route-controller-manager-576968d44d-lgnz7" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.276054 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6de61e80-75b6-402a-8166-65555cd3a0f7-client-ca\") pod \"route-controller-manager-576968d44d-lgnz7\" (UID: \"6de61e80-75b6-402a-8166-65555cd3a0f7\") " pod="openshift-route-controller-manager/route-controller-manager-576968d44d-lgnz7" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.275060 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6de61e80-75b6-402a-8166-65555cd3a0f7-config\") pod \"route-controller-manager-576968d44d-lgnz7\" (UID: \"6de61e80-75b6-402a-8166-65555cd3a0f7\") " pod="openshift-route-controller-manager/route-controller-manager-576968d44d-lgnz7" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.276143 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6de61e80-75b6-402a-8166-65555cd3a0f7-serving-cert\") pod \"route-controller-manager-576968d44d-lgnz7\" (UID: \"6de61e80-75b6-402a-8166-65555cd3a0f7\") " pod="openshift-route-controller-manager/route-controller-manager-576968d44d-lgnz7" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.276341 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6de61e80-75b6-402a-8166-65555cd3a0f7-config\") pod \"route-controller-manager-576968d44d-lgnz7\" (UID: \"6de61e80-75b6-402a-8166-65555cd3a0f7\") " pod="openshift-route-controller-manager/route-controller-manager-576968d44d-lgnz7" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.279228 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6de61e80-75b6-402a-8166-65555cd3a0f7-serving-cert\") pod \"route-controller-manager-576968d44d-lgnz7\" (UID: \"6de61e80-75b6-402a-8166-65555cd3a0f7\") " pod="openshift-route-controller-manager/route-controller-manager-576968d44d-lgnz7" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.291172 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4cds\" (UniqueName: \"kubernetes.io/projected/6de61e80-75b6-402a-8166-65555cd3a0f7-kube-api-access-c4cds\") pod \"route-controller-manager-576968d44d-lgnz7\" (UID: \"6de61e80-75b6-402a-8166-65555cd3a0f7\") " pod="openshift-route-controller-manager/route-controller-manager-576968d44d-lgnz7" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.430630 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f69f9958d-lwmg4" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.436624 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-576968d44d-lgnz7" Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.672159 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-576968d44d-lgnz7"] Feb 18 15:00:12 crc kubenswrapper[4896]: W0218 15:00:12.677335 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6de61e80_75b6_402a_8166_65555cd3a0f7.slice/crio-6ffef315b07fa931e1686a7032e5715990901cec9d57e6f96376c1b5b6549024 WatchSource:0}: Error finding container 6ffef315b07fa931e1686a7032e5715990901cec9d57e6f96376c1b5b6549024: Status 404 returned error can't find the container with id 6ffef315b07fa931e1686a7032e5715990901cec9d57e6f96376c1b5b6549024 Feb 18 15:00:12 crc kubenswrapper[4896]: I0218 15:00:12.821367 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5f69f9958d-lwmg4"] Feb 18 15:00:13 crc kubenswrapper[4896]: I0218 15:00:13.626609 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-576968d44d-lgnz7" event={"ID":"6de61e80-75b6-402a-8166-65555cd3a0f7","Type":"ContainerStarted","Data":"a54f63e03e5cf2a5db0733941cf0e22bf14d48388638e28459bdef14f66b638f"} Feb 18 15:00:13 crc kubenswrapper[4896]: I0218 15:00:13.626914 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-576968d44d-lgnz7" event={"ID":"6de61e80-75b6-402a-8166-65555cd3a0f7","Type":"ContainerStarted","Data":"6ffef315b07fa931e1686a7032e5715990901cec9d57e6f96376c1b5b6549024"} Feb 18 15:00:13 crc kubenswrapper[4896]: I0218 15:00:13.628323 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f69f9958d-lwmg4" event={"ID":"e8f7e30b-d57f-4849-8750-bc501da6b0f6","Type":"ContainerStarted","Data":"b0b1ec4839b511f03feefee0eb7483e4a41fb1775d596386125325407f3ecb66"} Feb 18 15:00:13 crc kubenswrapper[4896]: I0218 15:00:13.628370 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f69f9958d-lwmg4" event={"ID":"e8f7e30b-d57f-4849-8750-bc501da6b0f6","Type":"ContainerStarted","Data":"8718ba9b8ce670a0e1e9a3bf763c628b83348ad036500c130fd5d3d3af777b2c"} Feb 18 15:00:13 crc kubenswrapper[4896]: I0218 15:00:13.628594 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5f69f9958d-lwmg4" Feb 18 15:00:13 crc kubenswrapper[4896]: I0218 15:00:13.633394 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5f69f9958d-lwmg4" Feb 18 15:00:13 crc kubenswrapper[4896]: I0218 15:00:13.649518 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-576968d44d-lgnz7" podStartSLOduration=3.649499268 podStartE2EDuration="3.649499268s" podCreationTimestamp="2026-02-18 15:00:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:00:13.647337118 +0000 UTC m=+312.576671331" watchObservedRunningTime="2026-02-18 15:00:13.649499268 +0000 UTC m=+312.578833481" Feb 18 15:00:13 crc kubenswrapper[4896]: I0218 15:00:13.663146 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5f69f9958d-lwmg4" podStartSLOduration=3.663130784 podStartE2EDuration="3.663130784s" podCreationTimestamp="2026-02-18 15:00:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:00:13.662122346 +0000 UTC m=+312.591456589" watchObservedRunningTime="2026-02-18 15:00:13.663130784 +0000 UTC m=+312.592464997" Feb 18 15:00:14 crc kubenswrapper[4896]: I0218 15:00:14.633278 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-576968d44d-lgnz7" Feb 18 15:00:14 crc kubenswrapper[4896]: I0218 15:00:14.643540 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-576968d44d-lgnz7" Feb 18 15:00:30 crc kubenswrapper[4896]: I0218 15:00:30.024315 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-xrr77"] Feb 18 15:00:30 crc kubenswrapper[4896]: I0218 15:00:30.025484 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-xrr77" Feb 18 15:00:30 crc kubenswrapper[4896]: I0218 15:00:30.036660 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-xrr77"] Feb 18 15:00:30 crc kubenswrapper[4896]: I0218 15:00:30.155699 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/56e65c00-605a-4cc9-be5f-88044787cd0f-trusted-ca\") pod \"image-registry-66df7c8f76-xrr77\" (UID: \"56e65c00-605a-4cc9-be5f-88044787cd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrr77" Feb 18 15:00:30 crc kubenswrapper[4896]: I0218 15:00:30.155744 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8knst\" (UniqueName: \"kubernetes.io/projected/56e65c00-605a-4cc9-be5f-88044787cd0f-kube-api-access-8knst\") pod \"image-registry-66df7c8f76-xrr77\" (UID: \"56e65c00-605a-4cc9-be5f-88044787cd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrr77" Feb 18 15:00:30 crc kubenswrapper[4896]: I0218 15:00:30.155770 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/56e65c00-605a-4cc9-be5f-88044787cd0f-installation-pull-secrets\") pod \"image-registry-66df7c8f76-xrr77\" (UID: \"56e65c00-605a-4cc9-be5f-88044787cd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrr77" Feb 18 15:00:30 crc kubenswrapper[4896]: I0218 15:00:30.155884 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/56e65c00-605a-4cc9-be5f-88044787cd0f-registry-tls\") pod \"image-registry-66df7c8f76-xrr77\" (UID: \"56e65c00-605a-4cc9-be5f-88044787cd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrr77" Feb 18 15:00:30 crc kubenswrapper[4896]: I0218 15:00:30.155939 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/56e65c00-605a-4cc9-be5f-88044787cd0f-ca-trust-extracted\") pod \"image-registry-66df7c8f76-xrr77\" (UID: \"56e65c00-605a-4cc9-be5f-88044787cd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrr77" Feb 18 15:00:30 crc kubenswrapper[4896]: I0218 15:00:30.156079 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/56e65c00-605a-4cc9-be5f-88044787cd0f-bound-sa-token\") pod \"image-registry-66df7c8f76-xrr77\" (UID: \"56e65c00-605a-4cc9-be5f-88044787cd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrr77" Feb 18 15:00:30 crc kubenswrapper[4896]: I0218 15:00:30.156185 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/56e65c00-605a-4cc9-be5f-88044787cd0f-registry-certificates\") pod \"image-registry-66df7c8f76-xrr77\" (UID: \"56e65c00-605a-4cc9-be5f-88044787cd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrr77" Feb 18 15:00:30 crc kubenswrapper[4896]: I0218 15:00:30.156284 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-xrr77\" (UID: \"56e65c00-605a-4cc9-be5f-88044787cd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrr77" Feb 18 15:00:30 crc kubenswrapper[4896]: I0218 15:00:30.182206 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-xrr77\" (UID: \"56e65c00-605a-4cc9-be5f-88044787cd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrr77" Feb 18 15:00:30 crc kubenswrapper[4896]: I0218 15:00:30.257370 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/56e65c00-605a-4cc9-be5f-88044787cd0f-bound-sa-token\") pod \"image-registry-66df7c8f76-xrr77\" (UID: \"56e65c00-605a-4cc9-be5f-88044787cd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrr77" Feb 18 15:00:30 crc kubenswrapper[4896]: I0218 15:00:30.257426 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/56e65c00-605a-4cc9-be5f-88044787cd0f-registry-certificates\") pod \"image-registry-66df7c8f76-xrr77\" (UID: \"56e65c00-605a-4cc9-be5f-88044787cd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrr77" Feb 18 15:00:30 crc kubenswrapper[4896]: I0218 15:00:30.257457 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/56e65c00-605a-4cc9-be5f-88044787cd0f-trusted-ca\") pod \"image-registry-66df7c8f76-xrr77\" (UID: \"56e65c00-605a-4cc9-be5f-88044787cd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrr77" Feb 18 15:00:30 crc kubenswrapper[4896]: I0218 15:00:30.257475 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8knst\" (UniqueName: \"kubernetes.io/projected/56e65c00-605a-4cc9-be5f-88044787cd0f-kube-api-access-8knst\") pod \"image-registry-66df7c8f76-xrr77\" (UID: \"56e65c00-605a-4cc9-be5f-88044787cd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrr77" Feb 18 15:00:30 crc kubenswrapper[4896]: I0218 15:00:30.257493 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/56e65c00-605a-4cc9-be5f-88044787cd0f-installation-pull-secrets\") pod \"image-registry-66df7c8f76-xrr77\" (UID: \"56e65c00-605a-4cc9-be5f-88044787cd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrr77" Feb 18 15:00:30 crc kubenswrapper[4896]: I0218 15:00:30.257508 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/56e65c00-605a-4cc9-be5f-88044787cd0f-registry-tls\") pod \"image-registry-66df7c8f76-xrr77\" (UID: \"56e65c00-605a-4cc9-be5f-88044787cd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrr77" Feb 18 15:00:30 crc kubenswrapper[4896]: I0218 15:00:30.257532 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/56e65c00-605a-4cc9-be5f-88044787cd0f-ca-trust-extracted\") pod \"image-registry-66df7c8f76-xrr77\" (UID: \"56e65c00-605a-4cc9-be5f-88044787cd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrr77" Feb 18 15:00:30 crc kubenswrapper[4896]: I0218 15:00:30.258226 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/56e65c00-605a-4cc9-be5f-88044787cd0f-ca-trust-extracted\") pod \"image-registry-66df7c8f76-xrr77\" (UID: \"56e65c00-605a-4cc9-be5f-88044787cd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrr77" Feb 18 15:00:30 crc kubenswrapper[4896]: I0218 15:00:30.259037 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/56e65c00-605a-4cc9-be5f-88044787cd0f-trusted-ca\") pod \"image-registry-66df7c8f76-xrr77\" (UID: \"56e65c00-605a-4cc9-be5f-88044787cd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrr77" Feb 18 15:00:30 crc kubenswrapper[4896]: I0218 15:00:30.260349 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/56e65c00-605a-4cc9-be5f-88044787cd0f-registry-certificates\") pod \"image-registry-66df7c8f76-xrr77\" (UID: \"56e65c00-605a-4cc9-be5f-88044787cd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrr77" Feb 18 15:00:30 crc kubenswrapper[4896]: I0218 15:00:30.262983 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/56e65c00-605a-4cc9-be5f-88044787cd0f-registry-tls\") pod \"image-registry-66df7c8f76-xrr77\" (UID: \"56e65c00-605a-4cc9-be5f-88044787cd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrr77" Feb 18 15:00:30 crc kubenswrapper[4896]: I0218 15:00:30.263335 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/56e65c00-605a-4cc9-be5f-88044787cd0f-installation-pull-secrets\") pod \"image-registry-66df7c8f76-xrr77\" (UID: \"56e65c00-605a-4cc9-be5f-88044787cd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrr77" Feb 18 15:00:30 crc kubenswrapper[4896]: I0218 15:00:30.271267 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/56e65c00-605a-4cc9-be5f-88044787cd0f-bound-sa-token\") pod \"image-registry-66df7c8f76-xrr77\" (UID: \"56e65c00-605a-4cc9-be5f-88044787cd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrr77" Feb 18 15:00:30 crc kubenswrapper[4896]: I0218 15:00:30.273445 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8knst\" (UniqueName: \"kubernetes.io/projected/56e65c00-605a-4cc9-be5f-88044787cd0f-kube-api-access-8knst\") pod \"image-registry-66df7c8f76-xrr77\" (UID: \"56e65c00-605a-4cc9-be5f-88044787cd0f\") " pod="openshift-image-registry/image-registry-66df7c8f76-xrr77" Feb 18 15:00:30 crc kubenswrapper[4896]: I0218 15:00:30.345057 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-xrr77" Feb 18 15:00:30 crc kubenswrapper[4896]: I0218 15:00:30.731665 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-xrr77"] Feb 18 15:00:30 crc kubenswrapper[4896]: W0218 15:00:30.736915 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56e65c00_605a_4cc9_be5f_88044787cd0f.slice/crio-faf82d841270a61ab971e3b09253e78ceaf6805bdea248cf3083b46db9e4a8a0 WatchSource:0}: Error finding container faf82d841270a61ab971e3b09253e78ceaf6805bdea248cf3083b46db9e4a8a0: Status 404 returned error can't find the container with id faf82d841270a61ab971e3b09253e78ceaf6805bdea248cf3083b46db9e4a8a0 Feb 18 15:00:31 crc kubenswrapper[4896]: I0218 15:00:31.708616 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-xrr77" event={"ID":"56e65c00-605a-4cc9-be5f-88044787cd0f","Type":"ContainerStarted","Data":"d6b364d21f7eff55fea0c69437d9b92ae6f7b4575989b60d11f70b1cb88b1b93"} Feb 18 15:00:31 crc kubenswrapper[4896]: I0218 15:00:31.708857 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-xrr77" Feb 18 15:00:31 crc kubenswrapper[4896]: I0218 15:00:31.708866 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-xrr77" event={"ID":"56e65c00-605a-4cc9-be5f-88044787cd0f","Type":"ContainerStarted","Data":"faf82d841270a61ab971e3b09253e78ceaf6805bdea248cf3083b46db9e4a8a0"} Feb 18 15:00:31 crc kubenswrapper[4896]: I0218 15:00:31.724582 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-xrr77" podStartSLOduration=1.724564618 podStartE2EDuration="1.724564618s" podCreationTimestamp="2026-02-18 15:00:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:00:31.723445119 +0000 UTC m=+330.652779332" watchObservedRunningTime="2026-02-18 15:00:31.724564618 +0000 UTC m=+330.653898831" Feb 18 15:00:50 crc kubenswrapper[4896]: I0218 15:00:50.351070 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-xrr77" Feb 18 15:00:50 crc kubenswrapper[4896]: I0218 15:00:50.396810 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qrwnx"] Feb 18 15:00:50 crc kubenswrapper[4896]: I0218 15:00:50.715552 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-46f87"] Feb 18 15:00:50 crc kubenswrapper[4896]: I0218 15:00:50.716164 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-46f87" podUID="aa56560c-aab8-481c-9e3e-ac97ed9b69a5" containerName="registry-server" containerID="cri-o://8f90a1e466747e8c2c4a90531097356a6c13422797c456afbab5da533704f60d" gracePeriod=30 Feb 18 15:00:50 crc kubenswrapper[4896]: I0218 15:00:50.722354 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cp8c2"] Feb 18 15:00:50 crc kubenswrapper[4896]: I0218 15:00:50.722677 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cp8c2" podUID="96fbf244-90e4-4ffb-8dab-1ad893364032" containerName="registry-server" containerID="cri-o://cfba7a7ca5cc789db0990395717770a98b6c2de038a044c1e687035aad716e5f" gracePeriod=30 Feb 18 15:00:50 crc kubenswrapper[4896]: I0218 15:00:50.736523 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xrhcj"] Feb 18 15:00:50 crc kubenswrapper[4896]: I0218 15:00:50.736727 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-xrhcj" podUID="0f95dce4-6d68-44ac-9dd3-215664bb1d75" containerName="marketplace-operator" containerID="cri-o://3ae5e1d63464b4ac3259ba7f756992398a78f3fd9bf860514b62bb5252ea5fb1" gracePeriod=30 Feb 18 15:00:50 crc kubenswrapper[4896]: I0218 15:00:50.752803 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cmxh5"] Feb 18 15:00:50 crc kubenswrapper[4896]: I0218 15:00:50.753085 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-cmxh5" podUID="d6bf362d-c626-4e5d-8ff4-6e847ef4eb47" containerName="registry-server" containerID="cri-o://0856168e13497c09e9b7ac5a43826f510095d802c443a5df89e2e149037e3508" gracePeriod=30 Feb 18 15:00:50 crc kubenswrapper[4896]: I0218 15:00:50.758099 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5s2vh"] Feb 18 15:00:50 crc kubenswrapper[4896]: I0218 15:00:50.758796 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5s2vh" Feb 18 15:00:50 crc kubenswrapper[4896]: I0218 15:00:50.770635 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gq2n2"] Feb 18 15:00:50 crc kubenswrapper[4896]: I0218 15:00:50.770918 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gq2n2" podUID="b8604b65-3d0c-40e3-93bb-00ff1169f8f9" containerName="registry-server" containerID="cri-o://4a7340c954072b2dbf2cb422d0aa88eed139f6a041257cbe8fb727ff319f0cb1" gracePeriod=30 Feb 18 15:00:50 crc kubenswrapper[4896]: I0218 15:00:50.776617 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5s2vh"] Feb 18 15:00:50 crc kubenswrapper[4896]: I0218 15:00:50.951362 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/18d24b18-a5d0-4fc8-899f-02bf11c3e16c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5s2vh\" (UID: \"18d24b18-a5d0-4fc8-899f-02bf11c3e16c\") " pod="openshift-marketplace/marketplace-operator-79b997595-5s2vh" Feb 18 15:00:50 crc kubenswrapper[4896]: I0218 15:00:50.951434 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/18d24b18-a5d0-4fc8-899f-02bf11c3e16c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5s2vh\" (UID: \"18d24b18-a5d0-4fc8-899f-02bf11c3e16c\") " pod="openshift-marketplace/marketplace-operator-79b997595-5s2vh" Feb 18 15:00:50 crc kubenswrapper[4896]: I0218 15:00:50.951481 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7cx4\" (UniqueName: \"kubernetes.io/projected/18d24b18-a5d0-4fc8-899f-02bf11c3e16c-kube-api-access-f7cx4\") pod \"marketplace-operator-79b997595-5s2vh\" (UID: \"18d24b18-a5d0-4fc8-899f-02bf11c3e16c\") " pod="openshift-marketplace/marketplace-operator-79b997595-5s2vh" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.052654 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/18d24b18-a5d0-4fc8-899f-02bf11c3e16c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5s2vh\" (UID: \"18d24b18-a5d0-4fc8-899f-02bf11c3e16c\") " pod="openshift-marketplace/marketplace-operator-79b997595-5s2vh" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.052732 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/18d24b18-a5d0-4fc8-899f-02bf11c3e16c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5s2vh\" (UID: \"18d24b18-a5d0-4fc8-899f-02bf11c3e16c\") " pod="openshift-marketplace/marketplace-operator-79b997595-5s2vh" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.052773 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7cx4\" (UniqueName: \"kubernetes.io/projected/18d24b18-a5d0-4fc8-899f-02bf11c3e16c-kube-api-access-f7cx4\") pod \"marketplace-operator-79b997595-5s2vh\" (UID: \"18d24b18-a5d0-4fc8-899f-02bf11c3e16c\") " pod="openshift-marketplace/marketplace-operator-79b997595-5s2vh" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.054224 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/18d24b18-a5d0-4fc8-899f-02bf11c3e16c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5s2vh\" (UID: \"18d24b18-a5d0-4fc8-899f-02bf11c3e16c\") " pod="openshift-marketplace/marketplace-operator-79b997595-5s2vh" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.058898 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/18d24b18-a5d0-4fc8-899f-02bf11c3e16c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5s2vh\" (UID: \"18d24b18-a5d0-4fc8-899f-02bf11c3e16c\") " pod="openshift-marketplace/marketplace-operator-79b997595-5s2vh" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.079520 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7cx4\" (UniqueName: \"kubernetes.io/projected/18d24b18-a5d0-4fc8-899f-02bf11c3e16c-kube-api-access-f7cx4\") pod \"marketplace-operator-79b997595-5s2vh\" (UID: \"18d24b18-a5d0-4fc8-899f-02bf11c3e16c\") " pod="openshift-marketplace/marketplace-operator-79b997595-5s2vh" Feb 18 15:00:51 crc kubenswrapper[4896]: E0218 15:00:51.192389 4896 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f90a1e466747e8c2c4a90531097356a6c13422797c456afbab5da533704f60d is running failed: container process not found" containerID="8f90a1e466747e8c2c4a90531097356a6c13422797c456afbab5da533704f60d" cmd=["grpc_health_probe","-addr=:50051"] Feb 18 15:00:51 crc kubenswrapper[4896]: E0218 15:00:51.204473 4896 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f90a1e466747e8c2c4a90531097356a6c13422797c456afbab5da533704f60d is running failed: container process not found" containerID="8f90a1e466747e8c2c4a90531097356a6c13422797c456afbab5da533704f60d" cmd=["grpc_health_probe","-addr=:50051"] Feb 18 15:00:51 crc kubenswrapper[4896]: E0218 15:00:51.208129 4896 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f90a1e466747e8c2c4a90531097356a6c13422797c456afbab5da533704f60d is running failed: container process not found" containerID="8f90a1e466747e8c2c4a90531097356a6c13422797c456afbab5da533704f60d" cmd=["grpc_health_probe","-addr=:50051"] Feb 18 15:00:51 crc kubenswrapper[4896]: E0218 15:00:51.208210 4896 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 8f90a1e466747e8c2c4a90531097356a6c13422797c456afbab5da533704f60d is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-46f87" podUID="aa56560c-aab8-481c-9e3e-ac97ed9b69a5" containerName="registry-server" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.225871 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5s2vh" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.234126 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cp8c2" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.274623 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gq2n2" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.276900 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xrhcj" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.312551 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-46f87" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.327042 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cmxh5" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.360209 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hsqjj\" (UniqueName: \"kubernetes.io/projected/96fbf244-90e4-4ffb-8dab-1ad893364032-kube-api-access-hsqjj\") pod \"96fbf244-90e4-4ffb-8dab-1ad893364032\" (UID: \"96fbf244-90e4-4ffb-8dab-1ad893364032\") " Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.360354 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96fbf244-90e4-4ffb-8dab-1ad893364032-catalog-content\") pod \"96fbf244-90e4-4ffb-8dab-1ad893364032\" (UID: \"96fbf244-90e4-4ffb-8dab-1ad893364032\") " Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.360405 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwt8v\" (UniqueName: \"kubernetes.io/projected/0f95dce4-6d68-44ac-9dd3-215664bb1d75-kube-api-access-lwt8v\") pod \"0f95dce4-6d68-44ac-9dd3-215664bb1d75\" (UID: \"0f95dce4-6d68-44ac-9dd3-215664bb1d75\") " Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.360481 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96fbf244-90e4-4ffb-8dab-1ad893364032-utilities\") pod \"96fbf244-90e4-4ffb-8dab-1ad893364032\" (UID: \"96fbf244-90e4-4ffb-8dab-1ad893364032\") " Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.360517 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0f95dce4-6d68-44ac-9dd3-215664bb1d75-marketplace-trusted-ca\") pod \"0f95dce4-6d68-44ac-9dd3-215664bb1d75\" (UID: \"0f95dce4-6d68-44ac-9dd3-215664bb1d75\") " Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.361859 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f95dce4-6d68-44ac-9dd3-215664bb1d75-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "0f95dce4-6d68-44ac-9dd3-215664bb1d75" (UID: "0f95dce4-6d68-44ac-9dd3-215664bb1d75"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.363150 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96fbf244-90e4-4ffb-8dab-1ad893364032-utilities" (OuterVolumeSpecName: "utilities") pod "96fbf244-90e4-4ffb-8dab-1ad893364032" (UID: "96fbf244-90e4-4ffb-8dab-1ad893364032"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.363770 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96fbf244-90e4-4ffb-8dab-1ad893364032-kube-api-access-hsqjj" (OuterVolumeSpecName: "kube-api-access-hsqjj") pod "96fbf244-90e4-4ffb-8dab-1ad893364032" (UID: "96fbf244-90e4-4ffb-8dab-1ad893364032"). InnerVolumeSpecName "kube-api-access-hsqjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.364792 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f95dce4-6d68-44ac-9dd3-215664bb1d75-kube-api-access-lwt8v" (OuterVolumeSpecName: "kube-api-access-lwt8v") pod "0f95dce4-6d68-44ac-9dd3-215664bb1d75" (UID: "0f95dce4-6d68-44ac-9dd3-215664bb1d75"). InnerVolumeSpecName "kube-api-access-lwt8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.413443 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96fbf244-90e4-4ffb-8dab-1ad893364032-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "96fbf244-90e4-4ffb-8dab-1ad893364032" (UID: "96fbf244-90e4-4ffb-8dab-1ad893364032"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.461802 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa56560c-aab8-481c-9e3e-ac97ed9b69a5-utilities\") pod \"aa56560c-aab8-481c-9e3e-ac97ed9b69a5\" (UID: \"aa56560c-aab8-481c-9e3e-ac97ed9b69a5\") " Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.462107 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8604b65-3d0c-40e3-93bb-00ff1169f8f9-catalog-content\") pod \"b8604b65-3d0c-40e3-93bb-00ff1169f8f9\" (UID: \"b8604b65-3d0c-40e3-93bb-00ff1169f8f9\") " Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.462146 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gv76m\" (UniqueName: \"kubernetes.io/projected/aa56560c-aab8-481c-9e3e-ac97ed9b69a5-kube-api-access-gv76m\") pod \"aa56560c-aab8-481c-9e3e-ac97ed9b69a5\" (UID: \"aa56560c-aab8-481c-9e3e-ac97ed9b69a5\") " Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.462172 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hd44s\" (UniqueName: \"kubernetes.io/projected/b8604b65-3d0c-40e3-93bb-00ff1169f8f9-kube-api-access-hd44s\") pod \"b8604b65-3d0c-40e3-93bb-00ff1169f8f9\" (UID: \"b8604b65-3d0c-40e3-93bb-00ff1169f8f9\") " Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.462291 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8604b65-3d0c-40e3-93bb-00ff1169f8f9-utilities\") pod \"b8604b65-3d0c-40e3-93bb-00ff1169f8f9\" (UID: \"b8604b65-3d0c-40e3-93bb-00ff1169f8f9\") " Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.462518 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6bf362d-c626-4e5d-8ff4-6e847ef4eb47-utilities\") pod \"d6bf362d-c626-4e5d-8ff4-6e847ef4eb47\" (UID: \"d6bf362d-c626-4e5d-8ff4-6e847ef4eb47\") " Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.462551 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6bf362d-c626-4e5d-8ff4-6e847ef4eb47-catalog-content\") pod \"d6bf362d-c626-4e5d-8ff4-6e847ef4eb47\" (UID: \"d6bf362d-c626-4e5d-8ff4-6e847ef4eb47\") " Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.462599 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa56560c-aab8-481c-9e3e-ac97ed9b69a5-catalog-content\") pod \"aa56560c-aab8-481c-9e3e-ac97ed9b69a5\" (UID: \"aa56560c-aab8-481c-9e3e-ac97ed9b69a5\") " Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.462620 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vtkk\" (UniqueName: \"kubernetes.io/projected/d6bf362d-c626-4e5d-8ff4-6e847ef4eb47-kube-api-access-8vtkk\") pod \"d6bf362d-c626-4e5d-8ff4-6e847ef4eb47\" (UID: \"d6bf362d-c626-4e5d-8ff4-6e847ef4eb47\") " Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.462652 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0f95dce4-6d68-44ac-9dd3-215664bb1d75-marketplace-operator-metrics\") pod \"0f95dce4-6d68-44ac-9dd3-215664bb1d75\" (UID: \"0f95dce4-6d68-44ac-9dd3-215664bb1d75\") " Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.464926 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6bf362d-c626-4e5d-8ff4-6e847ef4eb47-utilities" (OuterVolumeSpecName: "utilities") pod "d6bf362d-c626-4e5d-8ff4-6e847ef4eb47" (UID: "d6bf362d-c626-4e5d-8ff4-6e847ef4eb47"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.465863 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa56560c-aab8-481c-9e3e-ac97ed9b69a5-kube-api-access-gv76m" (OuterVolumeSpecName: "kube-api-access-gv76m") pod "aa56560c-aab8-481c-9e3e-ac97ed9b69a5" (UID: "aa56560c-aab8-481c-9e3e-ac97ed9b69a5"). InnerVolumeSpecName "kube-api-access-gv76m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.466264 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f95dce4-6d68-44ac-9dd3-215664bb1d75-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "0f95dce4-6d68-44ac-9dd3-215664bb1d75" (UID: "0f95dce4-6d68-44ac-9dd3-215664bb1d75"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.466886 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8604b65-3d0c-40e3-93bb-00ff1169f8f9-kube-api-access-hd44s" (OuterVolumeSpecName: "kube-api-access-hd44s") pod "b8604b65-3d0c-40e3-93bb-00ff1169f8f9" (UID: "b8604b65-3d0c-40e3-93bb-00ff1169f8f9"). InnerVolumeSpecName "kube-api-access-hd44s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.469503 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6bf362d-c626-4e5d-8ff4-6e847ef4eb47-kube-api-access-8vtkk" (OuterVolumeSpecName: "kube-api-access-8vtkk") pod "d6bf362d-c626-4e5d-8ff4-6e847ef4eb47" (UID: "d6bf362d-c626-4e5d-8ff4-6e847ef4eb47"). InnerVolumeSpecName "kube-api-access-8vtkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.469777 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa56560c-aab8-481c-9e3e-ac97ed9b69a5-utilities" (OuterVolumeSpecName: "utilities") pod "aa56560c-aab8-481c-9e3e-ac97ed9b69a5" (UID: "aa56560c-aab8-481c-9e3e-ac97ed9b69a5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.476350 4896 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d6bf362d-c626-4e5d-8ff4-6e847ef4eb47-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.476572 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hsqjj\" (UniqueName: \"kubernetes.io/projected/96fbf244-90e4-4ffb-8dab-1ad893364032-kube-api-access-hsqjj\") on node \"crc\" DevicePath \"\"" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.476596 4896 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0f95dce4-6d68-44ac-9dd3-215664bb1d75-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.476617 4896 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96fbf244-90e4-4ffb-8dab-1ad893364032-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.476835 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwt8v\" (UniqueName: \"kubernetes.io/projected/0f95dce4-6d68-44ac-9dd3-215664bb1d75-kube-api-access-lwt8v\") on node \"crc\" DevicePath \"\"" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.476849 4896 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96fbf244-90e4-4ffb-8dab-1ad893364032-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.476868 4896 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0f95dce4-6d68-44ac-9dd3-215664bb1d75-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.477078 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gv76m\" (UniqueName: \"kubernetes.io/projected/aa56560c-aab8-481c-9e3e-ac97ed9b69a5-kube-api-access-gv76m\") on node \"crc\" DevicePath \"\"" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.477097 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hd44s\" (UniqueName: \"kubernetes.io/projected/b8604b65-3d0c-40e3-93bb-00ff1169f8f9-kube-api-access-hd44s\") on node \"crc\" DevicePath \"\"" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.479376 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8604b65-3d0c-40e3-93bb-00ff1169f8f9-utilities" (OuterVolumeSpecName: "utilities") pod "b8604b65-3d0c-40e3-93bb-00ff1169f8f9" (UID: "b8604b65-3d0c-40e3-93bb-00ff1169f8f9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.497370 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6bf362d-c626-4e5d-8ff4-6e847ef4eb47-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d6bf362d-c626-4e5d-8ff4-6e847ef4eb47" (UID: "d6bf362d-c626-4e5d-8ff4-6e847ef4eb47"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.534506 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa56560c-aab8-481c-9e3e-ac97ed9b69a5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aa56560c-aab8-481c-9e3e-ac97ed9b69a5" (UID: "aa56560c-aab8-481c-9e3e-ac97ed9b69a5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.578165 4896 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa56560c-aab8-481c-9e3e-ac97ed9b69a5-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.578207 4896 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b8604b65-3d0c-40e3-93bb-00ff1169f8f9-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.578218 4896 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d6bf362d-c626-4e5d-8ff4-6e847ef4eb47-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.578247 4896 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa56560c-aab8-481c-9e3e-ac97ed9b69a5-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.578260 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vtkk\" (UniqueName: \"kubernetes.io/projected/d6bf362d-c626-4e5d-8ff4-6e847ef4eb47-kube-api-access-8vtkk\") on node \"crc\" DevicePath \"\"" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.601479 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8604b65-3d0c-40e3-93bb-00ff1169f8f9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b8604b65-3d0c-40e3-93bb-00ff1169f8f9" (UID: "b8604b65-3d0c-40e3-93bb-00ff1169f8f9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.649273 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5s2vh"] Feb 18 15:00:51 crc kubenswrapper[4896]: W0218 15:00:51.651530 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18d24b18_a5d0_4fc8_899f_02bf11c3e16c.slice/crio-597fc4ab9ffc0c3fdab8ea7cc6072c0c61ba145c68d71951f93209000afc436b WatchSource:0}: Error finding container 597fc4ab9ffc0c3fdab8ea7cc6072c0c61ba145c68d71951f93209000afc436b: Status 404 returned error can't find the container with id 597fc4ab9ffc0c3fdab8ea7cc6072c0c61ba145c68d71951f93209000afc436b Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.681500 4896 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b8604b65-3d0c-40e3-93bb-00ff1169f8f9-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.811139 4896 generic.go:334] "Generic (PLEG): container finished" podID="aa56560c-aab8-481c-9e3e-ac97ed9b69a5" containerID="8f90a1e466747e8c2c4a90531097356a6c13422797c456afbab5da533704f60d" exitCode=0 Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.811418 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-46f87" event={"ID":"aa56560c-aab8-481c-9e3e-ac97ed9b69a5","Type":"ContainerDied","Data":"8f90a1e466747e8c2c4a90531097356a6c13422797c456afbab5da533704f60d"} Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.811488 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-46f87" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.811533 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-46f87" event={"ID":"aa56560c-aab8-481c-9e3e-ac97ed9b69a5","Type":"ContainerDied","Data":"6ecce065ca1d1a22894bf23af68f120122311d31897b9422cd286634dfbac336"} Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.811559 4896 scope.go:117] "RemoveContainer" containerID="8f90a1e466747e8c2c4a90531097356a6c13422797c456afbab5da533704f60d" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.813978 4896 generic.go:334] "Generic (PLEG): container finished" podID="96fbf244-90e4-4ffb-8dab-1ad893364032" containerID="cfba7a7ca5cc789db0990395717770a98b6c2de038a044c1e687035aad716e5f" exitCode=0 Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.814042 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cp8c2" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.814253 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cp8c2" event={"ID":"96fbf244-90e4-4ffb-8dab-1ad893364032","Type":"ContainerDied","Data":"cfba7a7ca5cc789db0990395717770a98b6c2de038a044c1e687035aad716e5f"} Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.814383 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cp8c2" event={"ID":"96fbf244-90e4-4ffb-8dab-1ad893364032","Type":"ContainerDied","Data":"3f785dfd2aa6592795013d4b2127b50188c6f501e4d2378ab05789f64d713f50"} Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.818651 4896 generic.go:334] "Generic (PLEG): container finished" podID="b8604b65-3d0c-40e3-93bb-00ff1169f8f9" containerID="4a7340c954072b2dbf2cb422d0aa88eed139f6a041257cbe8fb727ff319f0cb1" exitCode=0 Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.818735 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gq2n2" event={"ID":"b8604b65-3d0c-40e3-93bb-00ff1169f8f9","Type":"ContainerDied","Data":"4a7340c954072b2dbf2cb422d0aa88eed139f6a041257cbe8fb727ff319f0cb1"} Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.819037 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gq2n2" event={"ID":"b8604b65-3d0c-40e3-93bb-00ff1169f8f9","Type":"ContainerDied","Data":"8b3d42e307e6afbebb469295c9fad75c7bdcb8d59d59061f6a7f0c26ad8de774"} Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.818754 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gq2n2" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.827980 4896 generic.go:334] "Generic (PLEG): container finished" podID="0f95dce4-6d68-44ac-9dd3-215664bb1d75" containerID="3ae5e1d63464b4ac3259ba7f756992398a78f3fd9bf860514b62bb5252ea5fb1" exitCode=0 Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.828068 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xrhcj" event={"ID":"0f95dce4-6d68-44ac-9dd3-215664bb1d75","Type":"ContainerDied","Data":"3ae5e1d63464b4ac3259ba7f756992398a78f3fd9bf860514b62bb5252ea5fb1"} Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.828097 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xrhcj" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.828119 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xrhcj" event={"ID":"0f95dce4-6d68-44ac-9dd3-215664bb1d75","Type":"ContainerDied","Data":"67c9b838de57980c73fed1e514d27d34433e345ed4f732650b1e366e7da115ca"} Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.830450 4896 generic.go:334] "Generic (PLEG): container finished" podID="d6bf362d-c626-4e5d-8ff4-6e847ef4eb47" containerID="0856168e13497c09e9b7ac5a43826f510095d802c443a5df89e2e149037e3508" exitCode=0 Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.830613 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cmxh5" event={"ID":"d6bf362d-c626-4e5d-8ff4-6e847ef4eb47","Type":"ContainerDied","Data":"0856168e13497c09e9b7ac5a43826f510095d802c443a5df89e2e149037e3508"} Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.830639 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cmxh5" event={"ID":"d6bf362d-c626-4e5d-8ff4-6e847ef4eb47","Type":"ContainerDied","Data":"214acc65ab155c5b608cc8d7946d0c061bdff7ef7c802d36193835c716eae06f"} Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.830716 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cmxh5" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.832787 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5s2vh" event={"ID":"18d24b18-a5d0-4fc8-899f-02bf11c3e16c","Type":"ContainerStarted","Data":"ece8cf8569e36d892050829106bebf754192d736a2dcab87360f79a6c42f249e"} Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.832824 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5s2vh" event={"ID":"18d24b18-a5d0-4fc8-899f-02bf11c3e16c","Type":"ContainerStarted","Data":"597fc4ab9ffc0c3fdab8ea7cc6072c0c61ba145c68d71951f93209000afc436b"} Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.833525 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-5s2vh" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.834590 4896 scope.go:117] "RemoveContainer" containerID="4c8d9c6a0e95bc6eb06861ee3753133baab9c1f1d84a35c7c6e5b1dc843dde32" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.834924 4896 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-5s2vh container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.66:8080/healthz\": dial tcp 10.217.0.66:8080: connect: connection refused" start-of-body= Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.834960 4896 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-5s2vh" podUID="18d24b18-a5d0-4fc8-899f-02bf11c3e16c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.66:8080/healthz\": dial tcp 10.217.0.66:8080: connect: connection refused" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.857852 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-5s2vh" podStartSLOduration=1.856983302 podStartE2EDuration="1.856983302s" podCreationTimestamp="2026-02-18 15:00:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:00:51.853784649 +0000 UTC m=+350.783118882" watchObservedRunningTime="2026-02-18 15:00:51.856983302 +0000 UTC m=+350.786317515" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.858897 4896 scope.go:117] "RemoveContainer" containerID="9b43efaeef2cc38cc545122e5ca4054ccb0509bf72f9d9341956d94094df65aa" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.922260 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-46f87"] Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.922301 4896 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-46f87"] Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.922493 4896 scope.go:117] "RemoveContainer" containerID="8f90a1e466747e8c2c4a90531097356a6c13422797c456afbab5da533704f60d" Feb 18 15:00:51 crc kubenswrapper[4896]: E0218 15:00:51.923084 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f90a1e466747e8c2c4a90531097356a6c13422797c456afbab5da533704f60d\": container with ID starting with 8f90a1e466747e8c2c4a90531097356a6c13422797c456afbab5da533704f60d not found: ID does not exist" containerID="8f90a1e466747e8c2c4a90531097356a6c13422797c456afbab5da533704f60d" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.923159 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f90a1e466747e8c2c4a90531097356a6c13422797c456afbab5da533704f60d"} err="failed to get container status \"8f90a1e466747e8c2c4a90531097356a6c13422797c456afbab5da533704f60d\": rpc error: code = NotFound desc = could not find container \"8f90a1e466747e8c2c4a90531097356a6c13422797c456afbab5da533704f60d\": container with ID starting with 8f90a1e466747e8c2c4a90531097356a6c13422797c456afbab5da533704f60d not found: ID does not exist" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.923193 4896 scope.go:117] "RemoveContainer" containerID="4c8d9c6a0e95bc6eb06861ee3753133baab9c1f1d84a35c7c6e5b1dc843dde32" Feb 18 15:00:51 crc kubenswrapper[4896]: E0218 15:00:51.924099 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c8d9c6a0e95bc6eb06861ee3753133baab9c1f1d84a35c7c6e5b1dc843dde32\": container with ID starting with 4c8d9c6a0e95bc6eb06861ee3753133baab9c1f1d84a35c7c6e5b1dc843dde32 not found: ID does not exist" containerID="4c8d9c6a0e95bc6eb06861ee3753133baab9c1f1d84a35c7c6e5b1dc843dde32" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.924129 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c8d9c6a0e95bc6eb06861ee3753133baab9c1f1d84a35c7c6e5b1dc843dde32"} err="failed to get container status \"4c8d9c6a0e95bc6eb06861ee3753133baab9c1f1d84a35c7c6e5b1dc843dde32\": rpc error: code = NotFound desc = could not find container \"4c8d9c6a0e95bc6eb06861ee3753133baab9c1f1d84a35c7c6e5b1dc843dde32\": container with ID starting with 4c8d9c6a0e95bc6eb06861ee3753133baab9c1f1d84a35c7c6e5b1dc843dde32 not found: ID does not exist" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.924143 4896 scope.go:117] "RemoveContainer" containerID="9b43efaeef2cc38cc545122e5ca4054ccb0509bf72f9d9341956d94094df65aa" Feb 18 15:00:51 crc kubenswrapper[4896]: E0218 15:00:51.924502 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b43efaeef2cc38cc545122e5ca4054ccb0509bf72f9d9341956d94094df65aa\": container with ID starting with 9b43efaeef2cc38cc545122e5ca4054ccb0509bf72f9d9341956d94094df65aa not found: ID does not exist" containerID="9b43efaeef2cc38cc545122e5ca4054ccb0509bf72f9d9341956d94094df65aa" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.924531 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b43efaeef2cc38cc545122e5ca4054ccb0509bf72f9d9341956d94094df65aa"} err="failed to get container status \"9b43efaeef2cc38cc545122e5ca4054ccb0509bf72f9d9341956d94094df65aa\": rpc error: code = NotFound desc = could not find container \"9b43efaeef2cc38cc545122e5ca4054ccb0509bf72f9d9341956d94094df65aa\": container with ID starting with 9b43efaeef2cc38cc545122e5ca4054ccb0509bf72f9d9341956d94094df65aa not found: ID does not exist" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.924575 4896 scope.go:117] "RemoveContainer" containerID="cfba7a7ca5cc789db0990395717770a98b6c2de038a044c1e687035aad716e5f" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.937728 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa56560c-aab8-481c-9e3e-ac97ed9b69a5" path="/var/lib/kubelet/pods/aa56560c-aab8-481c-9e3e-ac97ed9b69a5/volumes" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.941849 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gq2n2"] Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.944983 4896 scope.go:117] "RemoveContainer" containerID="41bc3f54497918a30d8c5eb72b232b41fb29216fe9fa56db8a9eb6c4394a1317" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.947705 4896 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gq2n2"] Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.956099 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xrhcj"] Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.965112 4896 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xrhcj"] Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.967373 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cp8c2"] Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.968981 4896 scope.go:117] "RemoveContainer" containerID="00c69b287acdca635a7fc79198fe36b93c612a0c32a7de997735ff7da5687b18" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.975887 4896 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cp8c2"] Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.986915 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cmxh5"] Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.987860 4896 scope.go:117] "RemoveContainer" containerID="cfba7a7ca5cc789db0990395717770a98b6c2de038a044c1e687035aad716e5f" Feb 18 15:00:51 crc kubenswrapper[4896]: E0218 15:00:51.988255 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfba7a7ca5cc789db0990395717770a98b6c2de038a044c1e687035aad716e5f\": container with ID starting with cfba7a7ca5cc789db0990395717770a98b6c2de038a044c1e687035aad716e5f not found: ID does not exist" containerID="cfba7a7ca5cc789db0990395717770a98b6c2de038a044c1e687035aad716e5f" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.988374 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfba7a7ca5cc789db0990395717770a98b6c2de038a044c1e687035aad716e5f"} err="failed to get container status \"cfba7a7ca5cc789db0990395717770a98b6c2de038a044c1e687035aad716e5f\": rpc error: code = NotFound desc = could not find container \"cfba7a7ca5cc789db0990395717770a98b6c2de038a044c1e687035aad716e5f\": container with ID starting with cfba7a7ca5cc789db0990395717770a98b6c2de038a044c1e687035aad716e5f not found: ID does not exist" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.988463 4896 scope.go:117] "RemoveContainer" containerID="41bc3f54497918a30d8c5eb72b232b41fb29216fe9fa56db8a9eb6c4394a1317" Feb 18 15:00:51 crc kubenswrapper[4896]: E0218 15:00:51.988826 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41bc3f54497918a30d8c5eb72b232b41fb29216fe9fa56db8a9eb6c4394a1317\": container with ID starting with 41bc3f54497918a30d8c5eb72b232b41fb29216fe9fa56db8a9eb6c4394a1317 not found: ID does not exist" containerID="41bc3f54497918a30d8c5eb72b232b41fb29216fe9fa56db8a9eb6c4394a1317" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.988873 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41bc3f54497918a30d8c5eb72b232b41fb29216fe9fa56db8a9eb6c4394a1317"} err="failed to get container status \"41bc3f54497918a30d8c5eb72b232b41fb29216fe9fa56db8a9eb6c4394a1317\": rpc error: code = NotFound desc = could not find container \"41bc3f54497918a30d8c5eb72b232b41fb29216fe9fa56db8a9eb6c4394a1317\": container with ID starting with 41bc3f54497918a30d8c5eb72b232b41fb29216fe9fa56db8a9eb6c4394a1317 not found: ID does not exist" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.988911 4896 scope.go:117] "RemoveContainer" containerID="00c69b287acdca635a7fc79198fe36b93c612a0c32a7de997735ff7da5687b18" Feb 18 15:00:51 crc kubenswrapper[4896]: E0218 15:00:51.989398 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00c69b287acdca635a7fc79198fe36b93c612a0c32a7de997735ff7da5687b18\": container with ID starting with 00c69b287acdca635a7fc79198fe36b93c612a0c32a7de997735ff7da5687b18 not found: ID does not exist" containerID="00c69b287acdca635a7fc79198fe36b93c612a0c32a7de997735ff7da5687b18" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.989459 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00c69b287acdca635a7fc79198fe36b93c612a0c32a7de997735ff7da5687b18"} err="failed to get container status \"00c69b287acdca635a7fc79198fe36b93c612a0c32a7de997735ff7da5687b18\": rpc error: code = NotFound desc = could not find container \"00c69b287acdca635a7fc79198fe36b93c612a0c32a7de997735ff7da5687b18\": container with ID starting with 00c69b287acdca635a7fc79198fe36b93c612a0c32a7de997735ff7da5687b18 not found: ID does not exist" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.989475 4896 scope.go:117] "RemoveContainer" containerID="4a7340c954072b2dbf2cb422d0aa88eed139f6a041257cbe8fb727ff319f0cb1" Feb 18 15:00:51 crc kubenswrapper[4896]: I0218 15:00:51.990026 4896 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-cmxh5"] Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.001374 4896 scope.go:117] "RemoveContainer" containerID="aa751a5090abfb9a77dadd6cfd8c664132c3aab4d7df2440826785ed07af0c26" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.020392 4896 scope.go:117] "RemoveContainer" containerID="2b66d9b2b9040933d5dc10a17b7f7c2b6f687fec249eb138255ce8952a4b29e5" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.034199 4896 scope.go:117] "RemoveContainer" containerID="4a7340c954072b2dbf2cb422d0aa88eed139f6a041257cbe8fb727ff319f0cb1" Feb 18 15:00:52 crc kubenswrapper[4896]: E0218 15:00:52.034642 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a7340c954072b2dbf2cb422d0aa88eed139f6a041257cbe8fb727ff319f0cb1\": container with ID starting with 4a7340c954072b2dbf2cb422d0aa88eed139f6a041257cbe8fb727ff319f0cb1 not found: ID does not exist" containerID="4a7340c954072b2dbf2cb422d0aa88eed139f6a041257cbe8fb727ff319f0cb1" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.034682 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a7340c954072b2dbf2cb422d0aa88eed139f6a041257cbe8fb727ff319f0cb1"} err="failed to get container status \"4a7340c954072b2dbf2cb422d0aa88eed139f6a041257cbe8fb727ff319f0cb1\": rpc error: code = NotFound desc = could not find container \"4a7340c954072b2dbf2cb422d0aa88eed139f6a041257cbe8fb727ff319f0cb1\": container with ID starting with 4a7340c954072b2dbf2cb422d0aa88eed139f6a041257cbe8fb727ff319f0cb1 not found: ID does not exist" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.034708 4896 scope.go:117] "RemoveContainer" containerID="aa751a5090abfb9a77dadd6cfd8c664132c3aab4d7df2440826785ed07af0c26" Feb 18 15:00:52 crc kubenswrapper[4896]: E0218 15:00:52.034993 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa751a5090abfb9a77dadd6cfd8c664132c3aab4d7df2440826785ed07af0c26\": container with ID starting with aa751a5090abfb9a77dadd6cfd8c664132c3aab4d7df2440826785ed07af0c26 not found: ID does not exist" containerID="aa751a5090abfb9a77dadd6cfd8c664132c3aab4d7df2440826785ed07af0c26" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.035039 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa751a5090abfb9a77dadd6cfd8c664132c3aab4d7df2440826785ed07af0c26"} err="failed to get container status \"aa751a5090abfb9a77dadd6cfd8c664132c3aab4d7df2440826785ed07af0c26\": rpc error: code = NotFound desc = could not find container \"aa751a5090abfb9a77dadd6cfd8c664132c3aab4d7df2440826785ed07af0c26\": container with ID starting with aa751a5090abfb9a77dadd6cfd8c664132c3aab4d7df2440826785ed07af0c26 not found: ID does not exist" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.035068 4896 scope.go:117] "RemoveContainer" containerID="2b66d9b2b9040933d5dc10a17b7f7c2b6f687fec249eb138255ce8952a4b29e5" Feb 18 15:00:52 crc kubenswrapper[4896]: E0218 15:00:52.035497 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b66d9b2b9040933d5dc10a17b7f7c2b6f687fec249eb138255ce8952a4b29e5\": container with ID starting with 2b66d9b2b9040933d5dc10a17b7f7c2b6f687fec249eb138255ce8952a4b29e5 not found: ID does not exist" containerID="2b66d9b2b9040933d5dc10a17b7f7c2b6f687fec249eb138255ce8952a4b29e5" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.035534 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b66d9b2b9040933d5dc10a17b7f7c2b6f687fec249eb138255ce8952a4b29e5"} err="failed to get container status \"2b66d9b2b9040933d5dc10a17b7f7c2b6f687fec249eb138255ce8952a4b29e5\": rpc error: code = NotFound desc = could not find container \"2b66d9b2b9040933d5dc10a17b7f7c2b6f687fec249eb138255ce8952a4b29e5\": container with ID starting with 2b66d9b2b9040933d5dc10a17b7f7c2b6f687fec249eb138255ce8952a4b29e5 not found: ID does not exist" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.035564 4896 scope.go:117] "RemoveContainer" containerID="3ae5e1d63464b4ac3259ba7f756992398a78f3fd9bf860514b62bb5252ea5fb1" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.051087 4896 scope.go:117] "RemoveContainer" containerID="95e1564d67b45208daa6c2d917ef9ddc045a644295fdcc1557bdf6c9fdbea897" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.063910 4896 scope.go:117] "RemoveContainer" containerID="3ae5e1d63464b4ac3259ba7f756992398a78f3fd9bf860514b62bb5252ea5fb1" Feb 18 15:00:52 crc kubenswrapper[4896]: E0218 15:00:52.064323 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ae5e1d63464b4ac3259ba7f756992398a78f3fd9bf860514b62bb5252ea5fb1\": container with ID starting with 3ae5e1d63464b4ac3259ba7f756992398a78f3fd9bf860514b62bb5252ea5fb1 not found: ID does not exist" containerID="3ae5e1d63464b4ac3259ba7f756992398a78f3fd9bf860514b62bb5252ea5fb1" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.064375 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ae5e1d63464b4ac3259ba7f756992398a78f3fd9bf860514b62bb5252ea5fb1"} err="failed to get container status \"3ae5e1d63464b4ac3259ba7f756992398a78f3fd9bf860514b62bb5252ea5fb1\": rpc error: code = NotFound desc = could not find container \"3ae5e1d63464b4ac3259ba7f756992398a78f3fd9bf860514b62bb5252ea5fb1\": container with ID starting with 3ae5e1d63464b4ac3259ba7f756992398a78f3fd9bf860514b62bb5252ea5fb1 not found: ID does not exist" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.064409 4896 scope.go:117] "RemoveContainer" containerID="95e1564d67b45208daa6c2d917ef9ddc045a644295fdcc1557bdf6c9fdbea897" Feb 18 15:00:52 crc kubenswrapper[4896]: E0218 15:00:52.064745 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95e1564d67b45208daa6c2d917ef9ddc045a644295fdcc1557bdf6c9fdbea897\": container with ID starting with 95e1564d67b45208daa6c2d917ef9ddc045a644295fdcc1557bdf6c9fdbea897 not found: ID does not exist" containerID="95e1564d67b45208daa6c2d917ef9ddc045a644295fdcc1557bdf6c9fdbea897" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.064769 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95e1564d67b45208daa6c2d917ef9ddc045a644295fdcc1557bdf6c9fdbea897"} err="failed to get container status \"95e1564d67b45208daa6c2d917ef9ddc045a644295fdcc1557bdf6c9fdbea897\": rpc error: code = NotFound desc = could not find container \"95e1564d67b45208daa6c2d917ef9ddc045a644295fdcc1557bdf6c9fdbea897\": container with ID starting with 95e1564d67b45208daa6c2d917ef9ddc045a644295fdcc1557bdf6c9fdbea897 not found: ID does not exist" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.064787 4896 scope.go:117] "RemoveContainer" containerID="0856168e13497c09e9b7ac5a43826f510095d802c443a5df89e2e149037e3508" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.081085 4896 scope.go:117] "RemoveContainer" containerID="87bc842934ed7f6a0934b6c5bc66530fc2f47ebc7203927496453fcd790d2ab4" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.096109 4896 scope.go:117] "RemoveContainer" containerID="722f7e38c7392b6614356182a05a0dccd0dfa1c4f827fcaf02d7d4e95ff0ca8d" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.107556 4896 scope.go:117] "RemoveContainer" containerID="0856168e13497c09e9b7ac5a43826f510095d802c443a5df89e2e149037e3508" Feb 18 15:00:52 crc kubenswrapper[4896]: E0218 15:00:52.108242 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0856168e13497c09e9b7ac5a43826f510095d802c443a5df89e2e149037e3508\": container with ID starting with 0856168e13497c09e9b7ac5a43826f510095d802c443a5df89e2e149037e3508 not found: ID does not exist" containerID="0856168e13497c09e9b7ac5a43826f510095d802c443a5df89e2e149037e3508" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.108303 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0856168e13497c09e9b7ac5a43826f510095d802c443a5df89e2e149037e3508"} err="failed to get container status \"0856168e13497c09e9b7ac5a43826f510095d802c443a5df89e2e149037e3508\": rpc error: code = NotFound desc = could not find container \"0856168e13497c09e9b7ac5a43826f510095d802c443a5df89e2e149037e3508\": container with ID starting with 0856168e13497c09e9b7ac5a43826f510095d802c443a5df89e2e149037e3508 not found: ID does not exist" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.108338 4896 scope.go:117] "RemoveContainer" containerID="87bc842934ed7f6a0934b6c5bc66530fc2f47ebc7203927496453fcd790d2ab4" Feb 18 15:00:52 crc kubenswrapper[4896]: E0218 15:00:52.108705 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87bc842934ed7f6a0934b6c5bc66530fc2f47ebc7203927496453fcd790d2ab4\": container with ID starting with 87bc842934ed7f6a0934b6c5bc66530fc2f47ebc7203927496453fcd790d2ab4 not found: ID does not exist" containerID="87bc842934ed7f6a0934b6c5bc66530fc2f47ebc7203927496453fcd790d2ab4" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.108738 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87bc842934ed7f6a0934b6c5bc66530fc2f47ebc7203927496453fcd790d2ab4"} err="failed to get container status \"87bc842934ed7f6a0934b6c5bc66530fc2f47ebc7203927496453fcd790d2ab4\": rpc error: code = NotFound desc = could not find container \"87bc842934ed7f6a0934b6c5bc66530fc2f47ebc7203927496453fcd790d2ab4\": container with ID starting with 87bc842934ed7f6a0934b6c5bc66530fc2f47ebc7203927496453fcd790d2ab4 not found: ID does not exist" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.108760 4896 scope.go:117] "RemoveContainer" containerID="722f7e38c7392b6614356182a05a0dccd0dfa1c4f827fcaf02d7d4e95ff0ca8d" Feb 18 15:00:52 crc kubenswrapper[4896]: E0218 15:00:52.109109 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"722f7e38c7392b6614356182a05a0dccd0dfa1c4f827fcaf02d7d4e95ff0ca8d\": container with ID starting with 722f7e38c7392b6614356182a05a0dccd0dfa1c4f827fcaf02d7d4e95ff0ca8d not found: ID does not exist" containerID="722f7e38c7392b6614356182a05a0dccd0dfa1c4f827fcaf02d7d4e95ff0ca8d" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.109167 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"722f7e38c7392b6614356182a05a0dccd0dfa1c4f827fcaf02d7d4e95ff0ca8d"} err="failed to get container status \"722f7e38c7392b6614356182a05a0dccd0dfa1c4f827fcaf02d7d4e95ff0ca8d\": rpc error: code = NotFound desc = could not find container \"722f7e38c7392b6614356182a05a0dccd0dfa1c4f827fcaf02d7d4e95ff0ca8d\": container with ID starting with 722f7e38c7392b6614356182a05a0dccd0dfa1c4f827fcaf02d7d4e95ff0ca8d not found: ID does not exist" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.850504 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-5s2vh" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.950764 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gds9w"] Feb 18 15:00:52 crc kubenswrapper[4896]: E0218 15:00:52.950990 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8604b65-3d0c-40e3-93bb-00ff1169f8f9" containerName="registry-server" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.951005 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8604b65-3d0c-40e3-93bb-00ff1169f8f9" containerName="registry-server" Feb 18 15:00:52 crc kubenswrapper[4896]: E0218 15:00:52.951016 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6bf362d-c626-4e5d-8ff4-6e847ef4eb47" containerName="registry-server" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.951024 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6bf362d-c626-4e5d-8ff4-6e847ef4eb47" containerName="registry-server" Feb 18 15:00:52 crc kubenswrapper[4896]: E0218 15:00:52.951037 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa56560c-aab8-481c-9e3e-ac97ed9b69a5" containerName="registry-server" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.951044 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa56560c-aab8-481c-9e3e-ac97ed9b69a5" containerName="registry-server" Feb 18 15:00:52 crc kubenswrapper[4896]: E0218 15:00:52.951054 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96fbf244-90e4-4ffb-8dab-1ad893364032" containerName="extract-utilities" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.951062 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="96fbf244-90e4-4ffb-8dab-1ad893364032" containerName="extract-utilities" Feb 18 15:00:52 crc kubenswrapper[4896]: E0218 15:00:52.951074 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8604b65-3d0c-40e3-93bb-00ff1169f8f9" containerName="extract-utilities" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.951081 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8604b65-3d0c-40e3-93bb-00ff1169f8f9" containerName="extract-utilities" Feb 18 15:00:52 crc kubenswrapper[4896]: E0218 15:00:52.951097 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f95dce4-6d68-44ac-9dd3-215664bb1d75" containerName="marketplace-operator" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.951104 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f95dce4-6d68-44ac-9dd3-215664bb1d75" containerName="marketplace-operator" Feb 18 15:00:52 crc kubenswrapper[4896]: E0218 15:00:52.951112 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96fbf244-90e4-4ffb-8dab-1ad893364032" containerName="extract-content" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.951119 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="96fbf244-90e4-4ffb-8dab-1ad893364032" containerName="extract-content" Feb 18 15:00:52 crc kubenswrapper[4896]: E0218 15:00:52.951129 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8604b65-3d0c-40e3-93bb-00ff1169f8f9" containerName="extract-content" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.951135 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8604b65-3d0c-40e3-93bb-00ff1169f8f9" containerName="extract-content" Feb 18 15:00:52 crc kubenswrapper[4896]: E0218 15:00:52.951143 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6bf362d-c626-4e5d-8ff4-6e847ef4eb47" containerName="extract-utilities" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.951151 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6bf362d-c626-4e5d-8ff4-6e847ef4eb47" containerName="extract-utilities" Feb 18 15:00:52 crc kubenswrapper[4896]: E0218 15:00:52.951160 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f95dce4-6d68-44ac-9dd3-215664bb1d75" containerName="marketplace-operator" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.951167 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f95dce4-6d68-44ac-9dd3-215664bb1d75" containerName="marketplace-operator" Feb 18 15:00:52 crc kubenswrapper[4896]: E0218 15:00:52.951177 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa56560c-aab8-481c-9e3e-ac97ed9b69a5" containerName="extract-content" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.951185 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa56560c-aab8-481c-9e3e-ac97ed9b69a5" containerName="extract-content" Feb 18 15:00:52 crc kubenswrapper[4896]: E0218 15:00:52.951196 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa56560c-aab8-481c-9e3e-ac97ed9b69a5" containerName="extract-utilities" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.951205 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa56560c-aab8-481c-9e3e-ac97ed9b69a5" containerName="extract-utilities" Feb 18 15:00:52 crc kubenswrapper[4896]: E0218 15:00:52.951214 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96fbf244-90e4-4ffb-8dab-1ad893364032" containerName="registry-server" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.951221 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="96fbf244-90e4-4ffb-8dab-1ad893364032" containerName="registry-server" Feb 18 15:00:52 crc kubenswrapper[4896]: E0218 15:00:52.951230 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6bf362d-c626-4e5d-8ff4-6e847ef4eb47" containerName="extract-content" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.951256 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6bf362d-c626-4e5d-8ff4-6e847ef4eb47" containerName="extract-content" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.951367 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8604b65-3d0c-40e3-93bb-00ff1169f8f9" containerName="registry-server" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.951378 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="96fbf244-90e4-4ffb-8dab-1ad893364032" containerName="registry-server" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.951387 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f95dce4-6d68-44ac-9dd3-215664bb1d75" containerName="marketplace-operator" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.951397 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6bf362d-c626-4e5d-8ff4-6e847ef4eb47" containerName="registry-server" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.951405 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f95dce4-6d68-44ac-9dd3-215664bb1d75" containerName="marketplace-operator" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.951417 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa56560c-aab8-481c-9e3e-ac97ed9b69a5" containerName="registry-server" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.952129 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gds9w" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.953894 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 18 15:00:52 crc kubenswrapper[4896]: I0218 15:00:52.961711 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gds9w"] Feb 18 15:00:53 crc kubenswrapper[4896]: I0218 15:00:53.099303 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13955048-63bd-4d03-8aa7-f05d5f69b4e2-utilities\") pod \"certified-operators-gds9w\" (UID: \"13955048-63bd-4d03-8aa7-f05d5f69b4e2\") " pod="openshift-marketplace/certified-operators-gds9w" Feb 18 15:00:53 crc kubenswrapper[4896]: I0218 15:00:53.099361 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcnbb\" (UniqueName: \"kubernetes.io/projected/13955048-63bd-4d03-8aa7-f05d5f69b4e2-kube-api-access-wcnbb\") pod \"certified-operators-gds9w\" (UID: \"13955048-63bd-4d03-8aa7-f05d5f69b4e2\") " pod="openshift-marketplace/certified-operators-gds9w" Feb 18 15:00:53 crc kubenswrapper[4896]: I0218 15:00:53.099387 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13955048-63bd-4d03-8aa7-f05d5f69b4e2-catalog-content\") pod \"certified-operators-gds9w\" (UID: \"13955048-63bd-4d03-8aa7-f05d5f69b4e2\") " pod="openshift-marketplace/certified-operators-gds9w" Feb 18 15:00:53 crc kubenswrapper[4896]: I0218 15:00:53.130587 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fdrnw"] Feb 18 15:00:53 crc kubenswrapper[4896]: I0218 15:00:53.131754 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fdrnw" Feb 18 15:00:53 crc kubenswrapper[4896]: I0218 15:00:53.133856 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 18 15:00:53 crc kubenswrapper[4896]: I0218 15:00:53.138319 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fdrnw"] Feb 18 15:00:53 crc kubenswrapper[4896]: I0218 15:00:53.200871 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13955048-63bd-4d03-8aa7-f05d5f69b4e2-utilities\") pod \"certified-operators-gds9w\" (UID: \"13955048-63bd-4d03-8aa7-f05d5f69b4e2\") " pod="openshift-marketplace/certified-operators-gds9w" Feb 18 15:00:53 crc kubenswrapper[4896]: I0218 15:00:53.201227 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcnbb\" (UniqueName: \"kubernetes.io/projected/13955048-63bd-4d03-8aa7-f05d5f69b4e2-kube-api-access-wcnbb\") pod \"certified-operators-gds9w\" (UID: \"13955048-63bd-4d03-8aa7-f05d5f69b4e2\") " pod="openshift-marketplace/certified-operators-gds9w" Feb 18 15:00:53 crc kubenswrapper[4896]: I0218 15:00:53.201339 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13955048-63bd-4d03-8aa7-f05d5f69b4e2-catalog-content\") pod \"certified-operators-gds9w\" (UID: \"13955048-63bd-4d03-8aa7-f05d5f69b4e2\") " pod="openshift-marketplace/certified-operators-gds9w" Feb 18 15:00:53 crc kubenswrapper[4896]: I0218 15:00:53.201391 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13955048-63bd-4d03-8aa7-f05d5f69b4e2-utilities\") pod \"certified-operators-gds9w\" (UID: \"13955048-63bd-4d03-8aa7-f05d5f69b4e2\") " pod="openshift-marketplace/certified-operators-gds9w" Feb 18 15:00:53 crc kubenswrapper[4896]: I0218 15:00:53.201705 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13955048-63bd-4d03-8aa7-f05d5f69b4e2-catalog-content\") pod \"certified-operators-gds9w\" (UID: \"13955048-63bd-4d03-8aa7-f05d5f69b4e2\") " pod="openshift-marketplace/certified-operators-gds9w" Feb 18 15:00:53 crc kubenswrapper[4896]: I0218 15:00:53.224218 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcnbb\" (UniqueName: \"kubernetes.io/projected/13955048-63bd-4d03-8aa7-f05d5f69b4e2-kube-api-access-wcnbb\") pod \"certified-operators-gds9w\" (UID: \"13955048-63bd-4d03-8aa7-f05d5f69b4e2\") " pod="openshift-marketplace/certified-operators-gds9w" Feb 18 15:00:53 crc kubenswrapper[4896]: I0218 15:00:53.268998 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gds9w" Feb 18 15:00:53 crc kubenswrapper[4896]: I0218 15:00:53.303014 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scz57\" (UniqueName: \"kubernetes.io/projected/354f5a56-d8d3-4fd8-bf60-ee5287896e6c-kube-api-access-scz57\") pod \"community-operators-fdrnw\" (UID: \"354f5a56-d8d3-4fd8-bf60-ee5287896e6c\") " pod="openshift-marketplace/community-operators-fdrnw" Feb 18 15:00:53 crc kubenswrapper[4896]: I0218 15:00:53.303085 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/354f5a56-d8d3-4fd8-bf60-ee5287896e6c-catalog-content\") pod \"community-operators-fdrnw\" (UID: \"354f5a56-d8d3-4fd8-bf60-ee5287896e6c\") " pod="openshift-marketplace/community-operators-fdrnw" Feb 18 15:00:53 crc kubenswrapper[4896]: I0218 15:00:53.303120 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/354f5a56-d8d3-4fd8-bf60-ee5287896e6c-utilities\") pod \"community-operators-fdrnw\" (UID: \"354f5a56-d8d3-4fd8-bf60-ee5287896e6c\") " pod="openshift-marketplace/community-operators-fdrnw" Feb 18 15:00:53 crc kubenswrapper[4896]: I0218 15:00:53.405300 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/354f5a56-d8d3-4fd8-bf60-ee5287896e6c-utilities\") pod \"community-operators-fdrnw\" (UID: \"354f5a56-d8d3-4fd8-bf60-ee5287896e6c\") " pod="openshift-marketplace/community-operators-fdrnw" Feb 18 15:00:53 crc kubenswrapper[4896]: I0218 15:00:53.405385 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scz57\" (UniqueName: \"kubernetes.io/projected/354f5a56-d8d3-4fd8-bf60-ee5287896e6c-kube-api-access-scz57\") pod \"community-operators-fdrnw\" (UID: \"354f5a56-d8d3-4fd8-bf60-ee5287896e6c\") " pod="openshift-marketplace/community-operators-fdrnw" Feb 18 15:00:53 crc kubenswrapper[4896]: I0218 15:00:53.405415 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/354f5a56-d8d3-4fd8-bf60-ee5287896e6c-catalog-content\") pod \"community-operators-fdrnw\" (UID: \"354f5a56-d8d3-4fd8-bf60-ee5287896e6c\") " pod="openshift-marketplace/community-operators-fdrnw" Feb 18 15:00:53 crc kubenswrapper[4896]: I0218 15:00:53.405863 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/354f5a56-d8d3-4fd8-bf60-ee5287896e6c-catalog-content\") pod \"community-operators-fdrnw\" (UID: \"354f5a56-d8d3-4fd8-bf60-ee5287896e6c\") " pod="openshift-marketplace/community-operators-fdrnw" Feb 18 15:00:53 crc kubenswrapper[4896]: I0218 15:00:53.406339 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/354f5a56-d8d3-4fd8-bf60-ee5287896e6c-utilities\") pod \"community-operators-fdrnw\" (UID: \"354f5a56-d8d3-4fd8-bf60-ee5287896e6c\") " pod="openshift-marketplace/community-operators-fdrnw" Feb 18 15:00:53 crc kubenswrapper[4896]: I0218 15:00:53.422908 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scz57\" (UniqueName: \"kubernetes.io/projected/354f5a56-d8d3-4fd8-bf60-ee5287896e6c-kube-api-access-scz57\") pod \"community-operators-fdrnw\" (UID: \"354f5a56-d8d3-4fd8-bf60-ee5287896e6c\") " pod="openshift-marketplace/community-operators-fdrnw" Feb 18 15:00:53 crc kubenswrapper[4896]: I0218 15:00:53.481830 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fdrnw" Feb 18 15:00:53 crc kubenswrapper[4896]: I0218 15:00:53.636749 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gds9w"] Feb 18 15:00:53 crc kubenswrapper[4896]: W0218 15:00:53.639143 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod13955048_63bd_4d03_8aa7_f05d5f69b4e2.slice/crio-389c5eef5bf85d7700a3f828e1b98d0696992fa6029d3ab5e46f45d5215dad3f WatchSource:0}: Error finding container 389c5eef5bf85d7700a3f828e1b98d0696992fa6029d3ab5e46f45d5215dad3f: Status 404 returned error can't find the container with id 389c5eef5bf85d7700a3f828e1b98d0696992fa6029d3ab5e46f45d5215dad3f Feb 18 15:00:53 crc kubenswrapper[4896]: I0218 15:00:53.843661 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fdrnw"] Feb 18 15:00:53 crc kubenswrapper[4896]: W0218 15:00:53.847969 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod354f5a56_d8d3_4fd8_bf60_ee5287896e6c.slice/crio-4744f4f7cbc253a339e5d11d56b7d61619b6c4f9bca08b38fb822658b2c7118f WatchSource:0}: Error finding container 4744f4f7cbc253a339e5d11d56b7d61619b6c4f9bca08b38fb822658b2c7118f: Status 404 returned error can't find the container with id 4744f4f7cbc253a339e5d11d56b7d61619b6c4f9bca08b38fb822658b2c7118f Feb 18 15:00:53 crc kubenswrapper[4896]: I0218 15:00:53.850455 4896 generic.go:334] "Generic (PLEG): container finished" podID="13955048-63bd-4d03-8aa7-f05d5f69b4e2" containerID="11a9749ddc7b3daa1349a6b358db5893c5602f0bb54a063966f5d1bc59cb9af0" exitCode=0 Feb 18 15:00:53 crc kubenswrapper[4896]: I0218 15:00:53.850552 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gds9w" event={"ID":"13955048-63bd-4d03-8aa7-f05d5f69b4e2","Type":"ContainerDied","Data":"11a9749ddc7b3daa1349a6b358db5893c5602f0bb54a063966f5d1bc59cb9af0"} Feb 18 15:00:53 crc kubenswrapper[4896]: I0218 15:00:53.850616 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gds9w" event={"ID":"13955048-63bd-4d03-8aa7-f05d5f69b4e2","Type":"ContainerStarted","Data":"389c5eef5bf85d7700a3f828e1b98d0696992fa6029d3ab5e46f45d5215dad3f"} Feb 18 15:00:53 crc kubenswrapper[4896]: I0218 15:00:53.939708 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f95dce4-6d68-44ac-9dd3-215664bb1d75" path="/var/lib/kubelet/pods/0f95dce4-6d68-44ac-9dd3-215664bb1d75/volumes" Feb 18 15:00:53 crc kubenswrapper[4896]: I0218 15:00:53.940568 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96fbf244-90e4-4ffb-8dab-1ad893364032" path="/var/lib/kubelet/pods/96fbf244-90e4-4ffb-8dab-1ad893364032/volumes" Feb 18 15:00:53 crc kubenswrapper[4896]: I0218 15:00:53.941357 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8604b65-3d0c-40e3-93bb-00ff1169f8f9" path="/var/lib/kubelet/pods/b8604b65-3d0c-40e3-93bb-00ff1169f8f9/volumes" Feb 18 15:00:53 crc kubenswrapper[4896]: I0218 15:00:53.942571 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6bf362d-c626-4e5d-8ff4-6e847ef4eb47" path="/var/lib/kubelet/pods/d6bf362d-c626-4e5d-8ff4-6e847ef4eb47/volumes" Feb 18 15:00:54 crc kubenswrapper[4896]: I0218 15:00:54.855666 4896 generic.go:334] "Generic (PLEG): container finished" podID="13955048-63bd-4d03-8aa7-f05d5f69b4e2" containerID="a902fe66ec6df44fb512ec1be0b853df84033a5c53de6a083b70ed724f955227" exitCode=0 Feb 18 15:00:54 crc kubenswrapper[4896]: I0218 15:00:54.855748 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gds9w" event={"ID":"13955048-63bd-4d03-8aa7-f05d5f69b4e2","Type":"ContainerDied","Data":"a902fe66ec6df44fb512ec1be0b853df84033a5c53de6a083b70ed724f955227"} Feb 18 15:00:54 crc kubenswrapper[4896]: I0218 15:00:54.858585 4896 generic.go:334] "Generic (PLEG): container finished" podID="354f5a56-d8d3-4fd8-bf60-ee5287896e6c" containerID="b51f567e1f2e49d24669fe4aa516c6244c86c372b8b13ff9ce11848ad8666afe" exitCode=0 Feb 18 15:00:54 crc kubenswrapper[4896]: I0218 15:00:54.858632 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fdrnw" event={"ID":"354f5a56-d8d3-4fd8-bf60-ee5287896e6c","Type":"ContainerDied","Data":"b51f567e1f2e49d24669fe4aa516c6244c86c372b8b13ff9ce11848ad8666afe"} Feb 18 15:00:54 crc kubenswrapper[4896]: I0218 15:00:54.858668 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fdrnw" event={"ID":"354f5a56-d8d3-4fd8-bf60-ee5287896e6c","Type":"ContainerStarted","Data":"4744f4f7cbc253a339e5d11d56b7d61619b6c4f9bca08b38fb822658b2c7118f"} Feb 18 15:00:55 crc kubenswrapper[4896]: I0218 15:00:55.333787 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-czqrg"] Feb 18 15:00:55 crc kubenswrapper[4896]: I0218 15:00:55.337914 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-czqrg" Feb 18 15:00:55 crc kubenswrapper[4896]: I0218 15:00:55.340994 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 18 15:00:55 crc kubenswrapper[4896]: I0218 15:00:55.344057 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-czqrg"] Feb 18 15:00:55 crc kubenswrapper[4896]: I0218 15:00:55.434059 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cd2aec2-d292-4074-80db-b06de2ce3118-catalog-content\") pod \"redhat-marketplace-czqrg\" (UID: \"8cd2aec2-d292-4074-80db-b06de2ce3118\") " pod="openshift-marketplace/redhat-marketplace-czqrg" Feb 18 15:00:55 crc kubenswrapper[4896]: I0218 15:00:55.434454 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cd2aec2-d292-4074-80db-b06de2ce3118-utilities\") pod \"redhat-marketplace-czqrg\" (UID: \"8cd2aec2-d292-4074-80db-b06de2ce3118\") " pod="openshift-marketplace/redhat-marketplace-czqrg" Feb 18 15:00:55 crc kubenswrapper[4896]: I0218 15:00:55.434504 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzx67\" (UniqueName: \"kubernetes.io/projected/8cd2aec2-d292-4074-80db-b06de2ce3118-kube-api-access-lzx67\") pod \"redhat-marketplace-czqrg\" (UID: \"8cd2aec2-d292-4074-80db-b06de2ce3118\") " pod="openshift-marketplace/redhat-marketplace-czqrg" Feb 18 15:00:55 crc kubenswrapper[4896]: I0218 15:00:55.531078 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qz6dd"] Feb 18 15:00:55 crc kubenswrapper[4896]: I0218 15:00:55.532586 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qz6dd" Feb 18 15:00:55 crc kubenswrapper[4896]: I0218 15:00:55.534606 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 18 15:00:55 crc kubenswrapper[4896]: I0218 15:00:55.534933 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cd2aec2-d292-4074-80db-b06de2ce3118-catalog-content\") pod \"redhat-marketplace-czqrg\" (UID: \"8cd2aec2-d292-4074-80db-b06de2ce3118\") " pod="openshift-marketplace/redhat-marketplace-czqrg" Feb 18 15:00:55 crc kubenswrapper[4896]: I0218 15:00:55.534956 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cd2aec2-d292-4074-80db-b06de2ce3118-utilities\") pod \"redhat-marketplace-czqrg\" (UID: \"8cd2aec2-d292-4074-80db-b06de2ce3118\") " pod="openshift-marketplace/redhat-marketplace-czqrg" Feb 18 15:00:55 crc kubenswrapper[4896]: I0218 15:00:55.535005 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9csw\" (UniqueName: \"kubernetes.io/projected/bae88343-fc19-473e-9f08-7be40a0ac90d-kube-api-access-r9csw\") pod \"redhat-operators-qz6dd\" (UID: \"bae88343-fc19-473e-9f08-7be40a0ac90d\") " pod="openshift-marketplace/redhat-operators-qz6dd" Feb 18 15:00:55 crc kubenswrapper[4896]: I0218 15:00:55.535026 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzx67\" (UniqueName: \"kubernetes.io/projected/8cd2aec2-d292-4074-80db-b06de2ce3118-kube-api-access-lzx67\") pod \"redhat-marketplace-czqrg\" (UID: \"8cd2aec2-d292-4074-80db-b06de2ce3118\") " pod="openshift-marketplace/redhat-marketplace-czqrg" Feb 18 15:00:55 crc kubenswrapper[4896]: I0218 15:00:55.535084 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bae88343-fc19-473e-9f08-7be40a0ac90d-catalog-content\") pod \"redhat-operators-qz6dd\" (UID: \"bae88343-fc19-473e-9f08-7be40a0ac90d\") " pod="openshift-marketplace/redhat-operators-qz6dd" Feb 18 15:00:55 crc kubenswrapper[4896]: I0218 15:00:55.535106 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bae88343-fc19-473e-9f08-7be40a0ac90d-utilities\") pod \"redhat-operators-qz6dd\" (UID: \"bae88343-fc19-473e-9f08-7be40a0ac90d\") " pod="openshift-marketplace/redhat-operators-qz6dd" Feb 18 15:00:55 crc kubenswrapper[4896]: I0218 15:00:55.535432 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8cd2aec2-d292-4074-80db-b06de2ce3118-utilities\") pod \"redhat-marketplace-czqrg\" (UID: \"8cd2aec2-d292-4074-80db-b06de2ce3118\") " pod="openshift-marketplace/redhat-marketplace-czqrg" Feb 18 15:00:55 crc kubenswrapper[4896]: I0218 15:00:55.535491 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8cd2aec2-d292-4074-80db-b06de2ce3118-catalog-content\") pod \"redhat-marketplace-czqrg\" (UID: \"8cd2aec2-d292-4074-80db-b06de2ce3118\") " pod="openshift-marketplace/redhat-marketplace-czqrg" Feb 18 15:00:55 crc kubenswrapper[4896]: I0218 15:00:55.542833 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qz6dd"] Feb 18 15:00:55 crc kubenswrapper[4896]: I0218 15:00:55.607613 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzx67\" (UniqueName: \"kubernetes.io/projected/8cd2aec2-d292-4074-80db-b06de2ce3118-kube-api-access-lzx67\") pod \"redhat-marketplace-czqrg\" (UID: \"8cd2aec2-d292-4074-80db-b06de2ce3118\") " pod="openshift-marketplace/redhat-marketplace-czqrg" Feb 18 15:00:55 crc kubenswrapper[4896]: I0218 15:00:55.636910 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9csw\" (UniqueName: \"kubernetes.io/projected/bae88343-fc19-473e-9f08-7be40a0ac90d-kube-api-access-r9csw\") pod \"redhat-operators-qz6dd\" (UID: \"bae88343-fc19-473e-9f08-7be40a0ac90d\") " pod="openshift-marketplace/redhat-operators-qz6dd" Feb 18 15:00:55 crc kubenswrapper[4896]: I0218 15:00:55.636978 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bae88343-fc19-473e-9f08-7be40a0ac90d-catalog-content\") pod \"redhat-operators-qz6dd\" (UID: \"bae88343-fc19-473e-9f08-7be40a0ac90d\") " pod="openshift-marketplace/redhat-operators-qz6dd" Feb 18 15:00:55 crc kubenswrapper[4896]: I0218 15:00:55.637002 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bae88343-fc19-473e-9f08-7be40a0ac90d-utilities\") pod \"redhat-operators-qz6dd\" (UID: \"bae88343-fc19-473e-9f08-7be40a0ac90d\") " pod="openshift-marketplace/redhat-operators-qz6dd" Feb 18 15:00:55 crc kubenswrapper[4896]: I0218 15:00:55.637409 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bae88343-fc19-473e-9f08-7be40a0ac90d-utilities\") pod \"redhat-operators-qz6dd\" (UID: \"bae88343-fc19-473e-9f08-7be40a0ac90d\") " pod="openshift-marketplace/redhat-operators-qz6dd" Feb 18 15:00:55 crc kubenswrapper[4896]: I0218 15:00:55.637732 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bae88343-fc19-473e-9f08-7be40a0ac90d-catalog-content\") pod \"redhat-operators-qz6dd\" (UID: \"bae88343-fc19-473e-9f08-7be40a0ac90d\") " pod="openshift-marketplace/redhat-operators-qz6dd" Feb 18 15:00:55 crc kubenswrapper[4896]: I0218 15:00:55.655118 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9csw\" (UniqueName: \"kubernetes.io/projected/bae88343-fc19-473e-9f08-7be40a0ac90d-kube-api-access-r9csw\") pod \"redhat-operators-qz6dd\" (UID: \"bae88343-fc19-473e-9f08-7be40a0ac90d\") " pod="openshift-marketplace/redhat-operators-qz6dd" Feb 18 15:00:55 crc kubenswrapper[4896]: I0218 15:00:55.704911 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-czqrg" Feb 18 15:00:55 crc kubenswrapper[4896]: I0218 15:00:55.866296 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gds9w" event={"ID":"13955048-63bd-4d03-8aa7-f05d5f69b4e2","Type":"ContainerStarted","Data":"43e924d54552fbfe54ec6a15f2275ac9c5a89daacc161e1c68748125b614e5a6"} Feb 18 15:00:55 crc kubenswrapper[4896]: I0218 15:00:55.871028 4896 generic.go:334] "Generic (PLEG): container finished" podID="354f5a56-d8d3-4fd8-bf60-ee5287896e6c" containerID="e50bd67a025db876625067d51ddf2198598fd1f62f7d99cdfcdc27ed7ba82879" exitCode=0 Feb 18 15:00:55 crc kubenswrapper[4896]: I0218 15:00:55.871069 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fdrnw" event={"ID":"354f5a56-d8d3-4fd8-bf60-ee5287896e6c","Type":"ContainerDied","Data":"e50bd67a025db876625067d51ddf2198598fd1f62f7d99cdfcdc27ed7ba82879"} Feb 18 15:00:55 crc kubenswrapper[4896]: I0218 15:00:55.885886 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gds9w" podStartSLOduration=2.407601536 podStartE2EDuration="3.885870224s" podCreationTimestamp="2026-02-18 15:00:52 +0000 UTC" firstStartedPulling="2026-02-18 15:00:53.851806027 +0000 UTC m=+352.781140240" lastFinishedPulling="2026-02-18 15:00:55.330074715 +0000 UTC m=+354.259408928" observedRunningTime="2026-02-18 15:00:55.8826101 +0000 UTC m=+354.811944313" watchObservedRunningTime="2026-02-18 15:00:55.885870224 +0000 UTC m=+354.815204437" Feb 18 15:00:55 crc kubenswrapper[4896]: I0218 15:00:55.890168 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qz6dd" Feb 18 15:00:56 crc kubenswrapper[4896]: I0218 15:00:56.095747 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-czqrg"] Feb 18 15:00:56 crc kubenswrapper[4896]: I0218 15:00:56.274200 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qz6dd"] Feb 18 15:00:56 crc kubenswrapper[4896]: W0218 15:00:56.305948 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbae88343_fc19_473e_9f08_7be40a0ac90d.slice/crio-3cf0d2bc15c3bc1411dbc5798cdc02b9bd4c3bcd02f43b3d873bdcb497b5b611 WatchSource:0}: Error finding container 3cf0d2bc15c3bc1411dbc5798cdc02b9bd4c3bcd02f43b3d873bdcb497b5b611: Status 404 returned error can't find the container with id 3cf0d2bc15c3bc1411dbc5798cdc02b9bd4c3bcd02f43b3d873bdcb497b5b611 Feb 18 15:00:56 crc kubenswrapper[4896]: I0218 15:00:56.878983 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fdrnw" event={"ID":"354f5a56-d8d3-4fd8-bf60-ee5287896e6c","Type":"ContainerStarted","Data":"99c2f8f0575490344d403d8fccbd7ebca152a5c1148c8b5215aac0c3473f4a06"} Feb 18 15:00:56 crc kubenswrapper[4896]: I0218 15:00:56.882075 4896 generic.go:334] "Generic (PLEG): container finished" podID="8cd2aec2-d292-4074-80db-b06de2ce3118" containerID="c5078f302653d164374423e9ad597d21fb310a856b6d1e16fe40edc6c425f884" exitCode=0 Feb 18 15:00:56 crc kubenswrapper[4896]: I0218 15:00:56.882143 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-czqrg" event={"ID":"8cd2aec2-d292-4074-80db-b06de2ce3118","Type":"ContainerDied","Data":"c5078f302653d164374423e9ad597d21fb310a856b6d1e16fe40edc6c425f884"} Feb 18 15:00:56 crc kubenswrapper[4896]: I0218 15:00:56.882165 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-czqrg" event={"ID":"8cd2aec2-d292-4074-80db-b06de2ce3118","Type":"ContainerStarted","Data":"78d9023fcdd84eab2ef225c1434776493b7c845dfad448ad7f1f1428707b0760"} Feb 18 15:00:56 crc kubenswrapper[4896]: I0218 15:00:56.886639 4896 generic.go:334] "Generic (PLEG): container finished" podID="bae88343-fc19-473e-9f08-7be40a0ac90d" containerID="12bf8532e18be34973d1d97283b2df2f3ac9465130efc4819f04b7018a8fdbf4" exitCode=0 Feb 18 15:00:56 crc kubenswrapper[4896]: I0218 15:00:56.886683 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qz6dd" event={"ID":"bae88343-fc19-473e-9f08-7be40a0ac90d","Type":"ContainerDied","Data":"12bf8532e18be34973d1d97283b2df2f3ac9465130efc4819f04b7018a8fdbf4"} Feb 18 15:00:56 crc kubenswrapper[4896]: I0218 15:00:56.886716 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qz6dd" event={"ID":"bae88343-fc19-473e-9f08-7be40a0ac90d","Type":"ContainerStarted","Data":"3cf0d2bc15c3bc1411dbc5798cdc02b9bd4c3bcd02f43b3d873bdcb497b5b611"} Feb 18 15:00:56 crc kubenswrapper[4896]: I0218 15:00:56.897465 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fdrnw" podStartSLOduration=2.233248069 podStartE2EDuration="3.897452397s" podCreationTimestamp="2026-02-18 15:00:53 +0000 UTC" firstStartedPulling="2026-02-18 15:00:54.860636499 +0000 UTC m=+353.789970712" lastFinishedPulling="2026-02-18 15:00:56.524840827 +0000 UTC m=+355.454175040" observedRunningTime="2026-02-18 15:00:56.895367803 +0000 UTC m=+355.824702016" watchObservedRunningTime="2026-02-18 15:00:56.897452397 +0000 UTC m=+355.826786610" Feb 18 15:00:57 crc kubenswrapper[4896]: I0218 15:00:57.894742 4896 generic.go:334] "Generic (PLEG): container finished" podID="8cd2aec2-d292-4074-80db-b06de2ce3118" containerID="7928e4d06528e810b0b9b9552ba6df9cd98557ef70d323d514ec72f6d1506399" exitCode=0 Feb 18 15:00:57 crc kubenswrapper[4896]: I0218 15:00:57.894847 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-czqrg" event={"ID":"8cd2aec2-d292-4074-80db-b06de2ce3118","Type":"ContainerDied","Data":"7928e4d06528e810b0b9b9552ba6df9cd98557ef70d323d514ec72f6d1506399"} Feb 18 15:00:58 crc kubenswrapper[4896]: I0218 15:00:58.909664 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-czqrg" event={"ID":"8cd2aec2-d292-4074-80db-b06de2ce3118","Type":"ContainerStarted","Data":"9c4148ab190da1280cc04338b1a4c5c9d4b451b2d071130a7eb6a83c55cf0ac0"} Feb 18 15:00:58 crc kubenswrapper[4896]: I0218 15:00:58.928548 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-czqrg" podStartSLOduration=2.354187635 podStartE2EDuration="3.928529898s" podCreationTimestamp="2026-02-18 15:00:55 +0000 UTC" firstStartedPulling="2026-02-18 15:00:56.883101637 +0000 UTC m=+355.812435850" lastFinishedPulling="2026-02-18 15:00:58.4574439 +0000 UTC m=+357.386778113" observedRunningTime="2026-02-18 15:00:58.925355706 +0000 UTC m=+357.854689919" watchObservedRunningTime="2026-02-18 15:00:58.928529898 +0000 UTC m=+357.857864111" Feb 18 15:00:59 crc kubenswrapper[4896]: I0218 15:00:59.916300 4896 generic.go:334] "Generic (PLEG): container finished" podID="bae88343-fc19-473e-9f08-7be40a0ac90d" containerID="00ee93d0dead9015eada78bda332e4fae57c3aa2b78ba7ffbdfa506753cd4828" exitCode=0 Feb 18 15:00:59 crc kubenswrapper[4896]: I0218 15:00:59.916376 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qz6dd" event={"ID":"bae88343-fc19-473e-9f08-7be40a0ac90d","Type":"ContainerDied","Data":"00ee93d0dead9015eada78bda332e4fae57c3aa2b78ba7ffbdfa506753cd4828"} Feb 18 15:01:01 crc kubenswrapper[4896]: I0218 15:01:01.934856 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qz6dd" event={"ID":"bae88343-fc19-473e-9f08-7be40a0ac90d","Type":"ContainerStarted","Data":"27014eb1a0dbace8115df4793760f75b6f5ce31aee24587e89fc7c2e7457999f"} Feb 18 15:01:01 crc kubenswrapper[4896]: I0218 15:01:01.950719 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qz6dd" podStartSLOduration=3.461848554 podStartE2EDuration="6.950700783s" podCreationTimestamp="2026-02-18 15:00:55 +0000 UTC" firstStartedPulling="2026-02-18 15:00:56.887733687 +0000 UTC m=+355.817067900" lastFinishedPulling="2026-02-18 15:01:00.376585916 +0000 UTC m=+359.305920129" observedRunningTime="2026-02-18 15:01:01.946902185 +0000 UTC m=+360.876236398" watchObservedRunningTime="2026-02-18 15:01:01.950700783 +0000 UTC m=+360.880034996" Feb 18 15:01:03 crc kubenswrapper[4896]: I0218 15:01:03.269497 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gds9w" Feb 18 15:01:03 crc kubenswrapper[4896]: I0218 15:01:03.271911 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gds9w" Feb 18 15:01:03 crc kubenswrapper[4896]: I0218 15:01:03.309710 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gds9w" Feb 18 15:01:03 crc kubenswrapper[4896]: I0218 15:01:03.482439 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fdrnw" Feb 18 15:01:03 crc kubenswrapper[4896]: I0218 15:01:03.482503 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fdrnw" Feb 18 15:01:03 crc kubenswrapper[4896]: I0218 15:01:03.521072 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fdrnw" Feb 18 15:01:03 crc kubenswrapper[4896]: I0218 15:01:03.983517 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fdrnw" Feb 18 15:01:03 crc kubenswrapper[4896]: I0218 15:01:03.985576 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gds9w" Feb 18 15:01:04 crc kubenswrapper[4896]: I0218 15:01:04.076109 4896 patch_prober.go:28] interesting pod/machine-config-daemon-24rq5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:01:04 crc kubenswrapper[4896]: I0218 15:01:04.076165 4896 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:01:05 crc kubenswrapper[4896]: I0218 15:01:05.705335 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-czqrg" Feb 18 15:01:05 crc kubenswrapper[4896]: I0218 15:01:05.705636 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-czqrg" Feb 18 15:01:05 crc kubenswrapper[4896]: I0218 15:01:05.752924 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-czqrg" Feb 18 15:01:05 crc kubenswrapper[4896]: I0218 15:01:05.890626 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qz6dd" Feb 18 15:01:05 crc kubenswrapper[4896]: I0218 15:01:05.890694 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qz6dd" Feb 18 15:01:05 crc kubenswrapper[4896]: I0218 15:01:05.991912 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-czqrg" Feb 18 15:01:06 crc kubenswrapper[4896]: I0218 15:01:06.934289 4896 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qz6dd" podUID="bae88343-fc19-473e-9f08-7be40a0ac90d" containerName="registry-server" probeResult="failure" output=< Feb 18 15:01:06 crc kubenswrapper[4896]: timeout: failed to connect service ":50051" within 1s Feb 18 15:01:06 crc kubenswrapper[4896]: > Feb 18 15:01:15 crc kubenswrapper[4896]: I0218 15:01:15.443746 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" podUID="351aa26e-939a-4b69-a7ee-55be7111ee6e" containerName="registry" containerID="cri-o://ab354663dabb4d91e61c157af9994bb2c779a26ba987a35acf4e43c3b4490abd" gracePeriod=30 Feb 18 15:01:15 crc kubenswrapper[4896]: I0218 15:01:15.864606 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 15:01:15 crc kubenswrapper[4896]: I0218 15:01:15.937853 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qz6dd" Feb 18 15:01:15 crc kubenswrapper[4896]: I0218 15:01:15.976950 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qz6dd" Feb 18 15:01:15 crc kubenswrapper[4896]: I0218 15:01:15.984745 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/351aa26e-939a-4b69-a7ee-55be7111ee6e-installation-pull-secrets\") pod \"351aa26e-939a-4b69-a7ee-55be7111ee6e\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " Feb 18 15:01:15 crc kubenswrapper[4896]: I0218 15:01:15.985020 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/351aa26e-939a-4b69-a7ee-55be7111ee6e-registry-tls\") pod \"351aa26e-939a-4b69-a7ee-55be7111ee6e\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " Feb 18 15:01:15 crc kubenswrapper[4896]: I0218 15:01:15.985142 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/351aa26e-939a-4b69-a7ee-55be7111ee6e-bound-sa-token\") pod \"351aa26e-939a-4b69-a7ee-55be7111ee6e\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " Feb 18 15:01:15 crc kubenswrapper[4896]: I0218 15:01:15.985274 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f88rx\" (UniqueName: \"kubernetes.io/projected/351aa26e-939a-4b69-a7ee-55be7111ee6e-kube-api-access-f88rx\") pod \"351aa26e-939a-4b69-a7ee-55be7111ee6e\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " Feb 18 15:01:15 crc kubenswrapper[4896]: I0218 15:01:15.985359 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/351aa26e-939a-4b69-a7ee-55be7111ee6e-trusted-ca\") pod \"351aa26e-939a-4b69-a7ee-55be7111ee6e\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " Feb 18 15:01:15 crc kubenswrapper[4896]: I0218 15:01:15.985469 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/351aa26e-939a-4b69-a7ee-55be7111ee6e-registry-certificates\") pod \"351aa26e-939a-4b69-a7ee-55be7111ee6e\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " Feb 18 15:01:15 crc kubenswrapper[4896]: I0218 15:01:15.985554 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/351aa26e-939a-4b69-a7ee-55be7111ee6e-ca-trust-extracted\") pod \"351aa26e-939a-4b69-a7ee-55be7111ee6e\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " Feb 18 15:01:15 crc kubenswrapper[4896]: I0218 15:01:15.985770 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"351aa26e-939a-4b69-a7ee-55be7111ee6e\" (UID: \"351aa26e-939a-4b69-a7ee-55be7111ee6e\") " Feb 18 15:01:15 crc kubenswrapper[4896]: I0218 15:01:15.986426 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/351aa26e-939a-4b69-a7ee-55be7111ee6e-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "351aa26e-939a-4b69-a7ee-55be7111ee6e" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:01:15 crc kubenswrapper[4896]: I0218 15:01:15.986441 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/351aa26e-939a-4b69-a7ee-55be7111ee6e-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "351aa26e-939a-4b69-a7ee-55be7111ee6e" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:01:15 crc kubenswrapper[4896]: I0218 15:01:15.988336 4896 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/351aa26e-939a-4b69-a7ee-55be7111ee6e-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 18 15:01:15 crc kubenswrapper[4896]: I0218 15:01:15.988449 4896 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/351aa26e-939a-4b69-a7ee-55be7111ee6e-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 18 15:01:15 crc kubenswrapper[4896]: I0218 15:01:15.991758 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/351aa26e-939a-4b69-a7ee-55be7111ee6e-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "351aa26e-939a-4b69-a7ee-55be7111ee6e" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:01:15 crc kubenswrapper[4896]: I0218 15:01:15.992894 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/351aa26e-939a-4b69-a7ee-55be7111ee6e-kube-api-access-f88rx" (OuterVolumeSpecName: "kube-api-access-f88rx") pod "351aa26e-939a-4b69-a7ee-55be7111ee6e" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e"). InnerVolumeSpecName "kube-api-access-f88rx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:01:15 crc kubenswrapper[4896]: I0218 15:01:15.993031 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/351aa26e-939a-4b69-a7ee-55be7111ee6e-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "351aa26e-939a-4b69-a7ee-55be7111ee6e" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:01:15 crc kubenswrapper[4896]: I0218 15:01:15.994720 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/351aa26e-939a-4b69-a7ee-55be7111ee6e-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "351aa26e-939a-4b69-a7ee-55be7111ee6e" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:01:15 crc kubenswrapper[4896]: I0218 15:01:15.999299 4896 generic.go:334] "Generic (PLEG): container finished" podID="351aa26e-939a-4b69-a7ee-55be7111ee6e" containerID="ab354663dabb4d91e61c157af9994bb2c779a26ba987a35acf4e43c3b4490abd" exitCode=0 Feb 18 15:01:15 crc kubenswrapper[4896]: I0218 15:01:15.999379 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "351aa26e-939a-4b69-a7ee-55be7111ee6e" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 18 15:01:15 crc kubenswrapper[4896]: I0218 15:01:15.999488 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" event={"ID":"351aa26e-939a-4b69-a7ee-55be7111ee6e","Type":"ContainerDied","Data":"ab354663dabb4d91e61c157af9994bb2c779a26ba987a35acf4e43c3b4490abd"} Feb 18 15:01:15 crc kubenswrapper[4896]: I0218 15:01:15.999560 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" event={"ID":"351aa26e-939a-4b69-a7ee-55be7111ee6e","Type":"ContainerDied","Data":"d0f8c759aecf10320a99df9dc665d78d9c1a89a4bab29001d2ab4eeaeb6b7807"} Feb 18 15:01:15 crc kubenswrapper[4896]: I0218 15:01:15.999583 4896 scope.go:117] "RemoveContainer" containerID="ab354663dabb4d91e61c157af9994bb2c779a26ba987a35acf4e43c3b4490abd" Feb 18 15:01:16 crc kubenswrapper[4896]: I0218 15:01:15.999793 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-qrwnx" Feb 18 15:01:16 crc kubenswrapper[4896]: I0218 15:01:16.008494 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/351aa26e-939a-4b69-a7ee-55be7111ee6e-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "351aa26e-939a-4b69-a7ee-55be7111ee6e" (UID: "351aa26e-939a-4b69-a7ee-55be7111ee6e"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:01:16 crc kubenswrapper[4896]: I0218 15:01:16.047418 4896 scope.go:117] "RemoveContainer" containerID="ab354663dabb4d91e61c157af9994bb2c779a26ba987a35acf4e43c3b4490abd" Feb 18 15:01:16 crc kubenswrapper[4896]: E0218 15:01:16.047919 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab354663dabb4d91e61c157af9994bb2c779a26ba987a35acf4e43c3b4490abd\": container with ID starting with ab354663dabb4d91e61c157af9994bb2c779a26ba987a35acf4e43c3b4490abd not found: ID does not exist" containerID="ab354663dabb4d91e61c157af9994bb2c779a26ba987a35acf4e43c3b4490abd" Feb 18 15:01:16 crc kubenswrapper[4896]: I0218 15:01:16.047954 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab354663dabb4d91e61c157af9994bb2c779a26ba987a35acf4e43c3b4490abd"} err="failed to get container status \"ab354663dabb4d91e61c157af9994bb2c779a26ba987a35acf4e43c3b4490abd\": rpc error: code = NotFound desc = could not find container \"ab354663dabb4d91e61c157af9994bb2c779a26ba987a35acf4e43c3b4490abd\": container with ID starting with ab354663dabb4d91e61c157af9994bb2c779a26ba987a35acf4e43c3b4490abd not found: ID does not exist" Feb 18 15:01:16 crc kubenswrapper[4896]: I0218 15:01:16.090129 4896 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/351aa26e-939a-4b69-a7ee-55be7111ee6e-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 18 15:01:16 crc kubenswrapper[4896]: I0218 15:01:16.090175 4896 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/351aa26e-939a-4b69-a7ee-55be7111ee6e-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 18 15:01:16 crc kubenswrapper[4896]: I0218 15:01:16.090186 4896 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/351aa26e-939a-4b69-a7ee-55be7111ee6e-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 18 15:01:16 crc kubenswrapper[4896]: I0218 15:01:16.090194 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f88rx\" (UniqueName: \"kubernetes.io/projected/351aa26e-939a-4b69-a7ee-55be7111ee6e-kube-api-access-f88rx\") on node \"crc\" DevicePath \"\"" Feb 18 15:01:16 crc kubenswrapper[4896]: I0218 15:01:16.090203 4896 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/351aa26e-939a-4b69-a7ee-55be7111ee6e-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 18 15:01:16 crc kubenswrapper[4896]: I0218 15:01:16.322749 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qrwnx"] Feb 18 15:01:16 crc kubenswrapper[4896]: I0218 15:01:16.330694 4896 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-qrwnx"] Feb 18 15:01:17 crc kubenswrapper[4896]: I0218 15:01:17.933762 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="351aa26e-939a-4b69-a7ee-55be7111ee6e" path="/var/lib/kubelet/pods/351aa26e-939a-4b69-a7ee-55be7111ee6e/volumes" Feb 18 15:01:34 crc kubenswrapper[4896]: I0218 15:01:34.075708 4896 patch_prober.go:28] interesting pod/machine-config-daemon-24rq5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:01:34 crc kubenswrapper[4896]: I0218 15:01:34.076230 4896 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:02:04 crc kubenswrapper[4896]: I0218 15:02:04.075927 4896 patch_prober.go:28] interesting pod/machine-config-daemon-24rq5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:02:04 crc kubenswrapper[4896]: I0218 15:02:04.076835 4896 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:02:04 crc kubenswrapper[4896]: I0218 15:02:04.076920 4896 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" Feb 18 15:02:04 crc kubenswrapper[4896]: I0218 15:02:04.078048 4896 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"aa2586e82cda0fb6f85fe0dbdbf47b3474eef31fa5931dd1eca7bed587048bb4"} pod="openshift-machine-config-operator/machine-config-daemon-24rq5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 18 15:02:04 crc kubenswrapper[4896]: I0218 15:02:04.078176 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" containerID="cri-o://aa2586e82cda0fb6f85fe0dbdbf47b3474eef31fa5931dd1eca7bed587048bb4" gracePeriod=600 Feb 18 15:02:04 crc kubenswrapper[4896]: I0218 15:02:04.249639 4896 generic.go:334] "Generic (PLEG): container finished" podID="8503734a-8111-4fab-beda-1f0f3b59615b" containerID="aa2586e82cda0fb6f85fe0dbdbf47b3474eef31fa5931dd1eca7bed587048bb4" exitCode=0 Feb 18 15:02:04 crc kubenswrapper[4896]: I0218 15:02:04.249728 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" event={"ID":"8503734a-8111-4fab-beda-1f0f3b59615b","Type":"ContainerDied","Data":"aa2586e82cda0fb6f85fe0dbdbf47b3474eef31fa5931dd1eca7bed587048bb4"} Feb 18 15:02:04 crc kubenswrapper[4896]: I0218 15:02:04.250065 4896 scope.go:117] "RemoveContainer" containerID="5dbbb8f9264d7f9d2bdfdbe78b53ed78f3e4194e05a657713de8d330f50619fc" Feb 18 15:02:05 crc kubenswrapper[4896]: I0218 15:02:05.258189 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" event={"ID":"8503734a-8111-4fab-beda-1f0f3b59615b","Type":"ContainerStarted","Data":"e75d0500266e6b7af87e9b2cc3d752ab85959f8c74e00e1817c50ea02f263e92"} Feb 18 15:04:04 crc kubenswrapper[4896]: I0218 15:04:04.075746 4896 patch_prober.go:28] interesting pod/machine-config-daemon-24rq5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:04:04 crc kubenswrapper[4896]: I0218 15:04:04.076309 4896 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:04:34 crc kubenswrapper[4896]: I0218 15:04:34.075763 4896 patch_prober.go:28] interesting pod/machine-config-daemon-24rq5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:04:34 crc kubenswrapper[4896]: I0218 15:04:34.076501 4896 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:05:04 crc kubenswrapper[4896]: I0218 15:05:04.075650 4896 patch_prober.go:28] interesting pod/machine-config-daemon-24rq5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:05:04 crc kubenswrapper[4896]: I0218 15:05:04.076103 4896 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:05:04 crc kubenswrapper[4896]: I0218 15:05:04.076142 4896 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" Feb 18 15:05:04 crc kubenswrapper[4896]: I0218 15:05:04.076665 4896 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e75d0500266e6b7af87e9b2cc3d752ab85959f8c74e00e1817c50ea02f263e92"} pod="openshift-machine-config-operator/machine-config-daemon-24rq5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 18 15:05:04 crc kubenswrapper[4896]: I0218 15:05:04.076723 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" containerID="cri-o://e75d0500266e6b7af87e9b2cc3d752ab85959f8c74e00e1817c50ea02f263e92" gracePeriod=600 Feb 18 15:05:04 crc kubenswrapper[4896]: I0218 15:05:04.270121 4896 generic.go:334] "Generic (PLEG): container finished" podID="8503734a-8111-4fab-beda-1f0f3b59615b" containerID="e75d0500266e6b7af87e9b2cc3d752ab85959f8c74e00e1817c50ea02f263e92" exitCode=0 Feb 18 15:05:04 crc kubenswrapper[4896]: I0218 15:05:04.270223 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" event={"ID":"8503734a-8111-4fab-beda-1f0f3b59615b","Type":"ContainerDied","Data":"e75d0500266e6b7af87e9b2cc3d752ab85959f8c74e00e1817c50ea02f263e92"} Feb 18 15:05:04 crc kubenswrapper[4896]: I0218 15:05:04.271133 4896 scope.go:117] "RemoveContainer" containerID="aa2586e82cda0fb6f85fe0dbdbf47b3474eef31fa5931dd1eca7bed587048bb4" Feb 18 15:05:05 crc kubenswrapper[4896]: I0218 15:05:05.279961 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" event={"ID":"8503734a-8111-4fab-beda-1f0f3b59615b","Type":"ContainerStarted","Data":"91d6b5d30beb14a038ea309b143bbc7367c3406254b6275cdfed44c037d67a19"} Feb 18 15:05:55 crc kubenswrapper[4896]: I0218 15:05:55.808865 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-w862c"] Feb 18 15:05:55 crc kubenswrapper[4896]: I0218 15:05:55.809907 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="ovn-controller" containerID="cri-o://22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf" gracePeriod=30 Feb 18 15:05:55 crc kubenswrapper[4896]: I0218 15:05:55.810040 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="northd" containerID="cri-o://1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc" gracePeriod=30 Feb 18 15:05:55 crc kubenswrapper[4896]: I0218 15:05:55.810102 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2" gracePeriod=30 Feb 18 15:05:55 crc kubenswrapper[4896]: I0218 15:05:55.810092 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="nbdb" containerID="cri-o://52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b" gracePeriod=30 Feb 18 15:05:55 crc kubenswrapper[4896]: I0218 15:05:55.810150 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="kube-rbac-proxy-node" containerID="cri-o://3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521" gracePeriod=30 Feb 18 15:05:55 crc kubenswrapper[4896]: I0218 15:05:55.810199 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="ovn-acl-logging" containerID="cri-o://185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a" gracePeriod=30 Feb 18 15:05:55 crc kubenswrapper[4896]: I0218 15:05:55.811329 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="sbdb" containerID="cri-o://d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad" gracePeriod=30 Feb 18 15:05:55 crc kubenswrapper[4896]: I0218 15:05:55.843732 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="ovnkube-controller" containerID="cri-o://4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30" gracePeriod=30 Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.122642 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w862c_449d40af-fe23-4350-b66f-de4ce6614177/ovnkube-controller/3.log" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.126083 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w862c_449d40af-fe23-4350-b66f-de4ce6614177/ovn-acl-logging/0.log" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.126678 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w862c_449d40af-fe23-4350-b66f-de4ce6614177/ovn-controller/0.log" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.127046 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.181756 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-xcq25"] Feb 18 15:05:56 crc kubenswrapper[4896]: E0218 15:05:56.181986 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="northd" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.182003 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="northd" Feb 18 15:05:56 crc kubenswrapper[4896]: E0218 15:05:56.182016 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="351aa26e-939a-4b69-a7ee-55be7111ee6e" containerName="registry" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.182023 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="351aa26e-939a-4b69-a7ee-55be7111ee6e" containerName="registry" Feb 18 15:05:56 crc kubenswrapper[4896]: E0218 15:05:56.182034 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="ovnkube-controller" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.182042 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="ovnkube-controller" Feb 18 15:05:56 crc kubenswrapper[4896]: E0218 15:05:56.182051 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="sbdb" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.182058 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="sbdb" Feb 18 15:05:56 crc kubenswrapper[4896]: E0218 15:05:56.182067 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="ovn-controller" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.182075 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="ovn-controller" Feb 18 15:05:56 crc kubenswrapper[4896]: E0218 15:05:56.182091 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="nbdb" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.182099 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="nbdb" Feb 18 15:05:56 crc kubenswrapper[4896]: E0218 15:05:56.182106 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="ovnkube-controller" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.182114 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="ovnkube-controller" Feb 18 15:05:56 crc kubenswrapper[4896]: E0218 15:05:56.182122 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="kubecfg-setup" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.182129 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="kubecfg-setup" Feb 18 15:05:56 crc kubenswrapper[4896]: E0218 15:05:56.182138 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="ovnkube-controller" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.182145 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="ovnkube-controller" Feb 18 15:05:56 crc kubenswrapper[4896]: E0218 15:05:56.182155 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="kube-rbac-proxy-ovn-metrics" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.182162 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="kube-rbac-proxy-ovn-metrics" Feb 18 15:05:56 crc kubenswrapper[4896]: E0218 15:05:56.182171 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="ovnkube-controller" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.182178 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="ovnkube-controller" Feb 18 15:05:56 crc kubenswrapper[4896]: E0218 15:05:56.182189 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="ovn-acl-logging" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.182196 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="ovn-acl-logging" Feb 18 15:05:56 crc kubenswrapper[4896]: E0218 15:05:56.182202 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="kube-rbac-proxy-node" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.182211 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="kube-rbac-proxy-node" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.182333 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="ovnkube-controller" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.182344 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="ovnkube-controller" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.182352 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="351aa26e-939a-4b69-a7ee-55be7111ee6e" containerName="registry" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.182363 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="kube-rbac-proxy-ovn-metrics" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.182372 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="ovnkube-controller" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.182380 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="kube-rbac-proxy-node" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.182389 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="nbdb" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.182398 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="sbdb" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.182407 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="ovn-controller" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.182417 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="northd" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.182425 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="ovn-acl-logging" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.182433 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="ovnkube-controller" Feb 18 15:05:56 crc kubenswrapper[4896]: E0218 15:05:56.182531 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="ovnkube-controller" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.182541 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="ovnkube-controller" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.182699 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="449d40af-fe23-4350-b66f-de4ce6614177" containerName="ovnkube-controller" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.184736 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.219940 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-var-lib-openvswitch\") pod \"449d40af-fe23-4350-b66f-de4ce6614177\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.219999 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-node-log\") pod \"449d40af-fe23-4350-b66f-de4ce6614177\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.220035 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/449d40af-fe23-4350-b66f-de4ce6614177-ovnkube-script-lib\") pod \"449d40af-fe23-4350-b66f-de4ce6614177\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.220063 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-run-systemd\") pod \"449d40af-fe23-4350-b66f-de4ce6614177\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.220074 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-run-openvswitch\") pod \"449d40af-fe23-4350-b66f-de4ce6614177\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.220285 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/449d40af-fe23-4350-b66f-de4ce6614177-ovn-node-metrics-cert\") pod \"449d40af-fe23-4350-b66f-de4ce6614177\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.220314 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-run-ovn\") pod \"449d40af-fe23-4350-b66f-de4ce6614177\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.220301 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "449d40af-fe23-4350-b66f-de4ce6614177" (UID: "449d40af-fe23-4350-b66f-de4ce6614177"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.220329 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-run-netns\") pod \"449d40af-fe23-4350-b66f-de4ce6614177\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.220382 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "449d40af-fe23-4350-b66f-de4ce6614177" (UID: "449d40af-fe23-4350-b66f-de4ce6614177"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.220403 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-slash\") pod \"449d40af-fe23-4350-b66f-de4ce6614177\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.220414 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "449d40af-fe23-4350-b66f-de4ce6614177" (UID: "449d40af-fe23-4350-b66f-de4ce6614177"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.220443 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-cni-netd\") pod \"449d40af-fe23-4350-b66f-de4ce6614177\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.220616 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zjjt\" (UniqueName: \"kubernetes.io/projected/449d40af-fe23-4350-b66f-de4ce6614177-kube-api-access-9zjjt\") pod \"449d40af-fe23-4350-b66f-de4ce6614177\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.220679 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/449d40af-fe23-4350-b66f-de4ce6614177-ovnkube-config\") pod \"449d40af-fe23-4350-b66f-de4ce6614177\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.220706 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-kubelet\") pod \"449d40af-fe23-4350-b66f-de4ce6614177\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.220746 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-var-lib-cni-networks-ovn-kubernetes\") pod \"449d40af-fe23-4350-b66f-de4ce6614177\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.220774 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/449d40af-fe23-4350-b66f-de4ce6614177-env-overrides\") pod \"449d40af-fe23-4350-b66f-de4ce6614177\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.220773 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/449d40af-fe23-4350-b66f-de4ce6614177-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "449d40af-fe23-4350-b66f-de4ce6614177" (UID: "449d40af-fe23-4350-b66f-de4ce6614177"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.220800 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-log-socket\") pod \"449d40af-fe23-4350-b66f-de4ce6614177\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.220827 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-log-socket" (OuterVolumeSpecName: "log-socket") pod "449d40af-fe23-4350-b66f-de4ce6614177" (UID: "449d40af-fe23-4350-b66f-de4ce6614177"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.220854 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-slash" (OuterVolumeSpecName: "host-slash") pod "449d40af-fe23-4350-b66f-de4ce6614177" (UID: "449d40af-fe23-4350-b66f-de4ce6614177"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.220863 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-etc-openvswitch\") pod \"449d40af-fe23-4350-b66f-de4ce6614177\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.220879 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "449d40af-fe23-4350-b66f-de4ce6614177" (UID: "449d40af-fe23-4350-b66f-de4ce6614177"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.220888 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-systemd-units\") pod \"449d40af-fe23-4350-b66f-de4ce6614177\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.220905 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-cni-bin\") pod \"449d40af-fe23-4350-b66f-de4ce6614177\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.220925 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-run-ovn-kubernetes\") pod \"449d40af-fe23-4350-b66f-de4ce6614177\" (UID: \"449d40af-fe23-4350-b66f-de4ce6614177\") " Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.221337 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-node-log" (OuterVolumeSpecName: "node-log") pod "449d40af-fe23-4350-b66f-de4ce6614177" (UID: "449d40af-fe23-4350-b66f-de4ce6614177"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.221411 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "449d40af-fe23-4350-b66f-de4ce6614177" (UID: "449d40af-fe23-4350-b66f-de4ce6614177"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.221439 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "449d40af-fe23-4350-b66f-de4ce6614177" (UID: "449d40af-fe23-4350-b66f-de4ce6614177"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.221362 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "449d40af-fe23-4350-b66f-de4ce6614177" (UID: "449d40af-fe23-4350-b66f-de4ce6614177"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.221449 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "449d40af-fe23-4350-b66f-de4ce6614177" (UID: "449d40af-fe23-4350-b66f-de4ce6614177"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.221481 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "449d40af-fe23-4350-b66f-de4ce6614177" (UID: "449d40af-fe23-4350-b66f-de4ce6614177"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.221505 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "449d40af-fe23-4350-b66f-de4ce6614177" (UID: "449d40af-fe23-4350-b66f-de4ce6614177"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.221577 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "449d40af-fe23-4350-b66f-de4ce6614177" (UID: "449d40af-fe23-4350-b66f-de4ce6614177"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.221671 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/449d40af-fe23-4350-b66f-de4ce6614177-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "449d40af-fe23-4350-b66f-de4ce6614177" (UID: "449d40af-fe23-4350-b66f-de4ce6614177"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.221750 4896 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-kubelet\") on node \"crc\" DevicePath \"\"" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.221769 4896 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.221782 4896 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-log-socket\") on node \"crc\" DevicePath \"\"" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.221791 4896 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.221800 4896 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-systemd-units\") on node \"crc\" DevicePath \"\"" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.221810 4896 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-cni-bin\") on node \"crc\" DevicePath \"\"" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.221819 4896 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.221828 4896 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.221837 4896 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-node-log\") on node \"crc\" DevicePath \"\"" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.221846 4896 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/449d40af-fe23-4350-b66f-de4ce6614177-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.221854 4896 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-run-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.221863 4896 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.221872 4896 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-run-netns\") on node \"crc\" DevicePath \"\"" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.221880 4896 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-slash\") on node \"crc\" DevicePath \"\"" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.221887 4896 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-host-cni-netd\") on node \"crc\" DevicePath \"\"" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.221964 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/449d40af-fe23-4350-b66f-de4ce6614177-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "449d40af-fe23-4350-b66f-de4ce6614177" (UID: "449d40af-fe23-4350-b66f-de4ce6614177"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.226365 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/449d40af-fe23-4350-b66f-de4ce6614177-kube-api-access-9zjjt" (OuterVolumeSpecName: "kube-api-access-9zjjt") pod "449d40af-fe23-4350-b66f-de4ce6614177" (UID: "449d40af-fe23-4350-b66f-de4ce6614177"). InnerVolumeSpecName "kube-api-access-9zjjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.226377 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/449d40af-fe23-4350-b66f-de4ce6614177-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "449d40af-fe23-4350-b66f-de4ce6614177" (UID: "449d40af-fe23-4350-b66f-de4ce6614177"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.232963 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "449d40af-fe23-4350-b66f-de4ce6614177" (UID: "449d40af-fe23-4350-b66f-de4ce6614177"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.322954 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-host-cni-netd\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.322999 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-etc-openvswitch\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.323018 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.323092 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-node-log\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.323113 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-host-slash\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.323132 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/958af1eb-f14f-4af1-9b53-543c7cd3fae9-ovnkube-script-lib\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.323154 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-host-run-ovn-kubernetes\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.323192 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-host-run-netns\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.323216 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-host-cni-bin\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.323275 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-run-openvswitch\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.323300 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/958af1eb-f14f-4af1-9b53-543c7cd3fae9-ovnkube-config\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.323322 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-host-kubelet\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.323347 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wchl\" (UniqueName: \"kubernetes.io/projected/958af1eb-f14f-4af1-9b53-543c7cd3fae9-kube-api-access-8wchl\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.323427 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-systemd-units\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.323500 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/958af1eb-f14f-4af1-9b53-543c7cd3fae9-env-overrides\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.323583 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-var-lib-openvswitch\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.323630 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/958af1eb-f14f-4af1-9b53-543c7cd3fae9-ovn-node-metrics-cert\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.323658 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-run-systemd\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.323687 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-run-ovn\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.323709 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-log-socket\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.323788 4896 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/449d40af-fe23-4350-b66f-de4ce6614177-run-systemd\") on node \"crc\" DevicePath \"\"" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.323806 4896 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/449d40af-fe23-4350-b66f-de4ce6614177-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.323818 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zjjt\" (UniqueName: \"kubernetes.io/projected/449d40af-fe23-4350-b66f-de4ce6614177-kube-api-access-9zjjt\") on node \"crc\" DevicePath \"\"" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.323826 4896 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/449d40af-fe23-4350-b66f-de4ce6614177-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.323836 4896 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/449d40af-fe23-4350-b66f-de4ce6614177-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.424920 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-node-log\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.424986 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-host-slash\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.425014 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/958af1eb-f14f-4af1-9b53-543c7cd3fae9-ovnkube-script-lib\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.425039 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-host-run-ovn-kubernetes\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.425055 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-host-run-netns\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.425079 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-host-cni-bin\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.425075 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-node-log\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.425098 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-run-openvswitch\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.425134 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-run-openvswitch\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.425154 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/958af1eb-f14f-4af1-9b53-543c7cd3fae9-ovnkube-config\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.425172 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-host-slash\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.425190 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-host-kubelet\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.425227 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wchl\" (UniqueName: \"kubernetes.io/projected/958af1eb-f14f-4af1-9b53-543c7cd3fae9-kube-api-access-8wchl\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.425266 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-systemd-units\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.425284 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/958af1eb-f14f-4af1-9b53-543c7cd3fae9-env-overrides\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.425313 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-var-lib-openvswitch\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.425343 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/958af1eb-f14f-4af1-9b53-543c7cd3fae9-ovn-node-metrics-cert\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.425374 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-run-systemd\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.425395 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-run-ovn\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.425414 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-log-socket\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.425444 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-host-cni-netd\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.425463 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-etc-openvswitch\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.425486 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.425602 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.425890 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/958af1eb-f14f-4af1-9b53-543c7cd3fae9-ovnkube-script-lib\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.425941 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-host-run-ovn-kubernetes\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.425967 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-host-run-netns\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.425997 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-host-cni-bin\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.426032 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-var-lib-openvswitch\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.426063 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-host-kubelet\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.426261 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-systemd-units\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.426273 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-run-ovn\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.426326 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-run-systemd\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.426383 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-host-cni-netd\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.426405 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-log-socket\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.426414 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/958af1eb-f14f-4af1-9b53-543c7cd3fae9-etc-openvswitch\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.426565 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/958af1eb-f14f-4af1-9b53-543c7cd3fae9-ovnkube-config\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.426761 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/958af1eb-f14f-4af1-9b53-543c7cd3fae9-env-overrides\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.429615 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/958af1eb-f14f-4af1-9b53-543c7cd3fae9-ovn-node-metrics-cert\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.440487 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wchl\" (UniqueName: \"kubernetes.io/projected/958af1eb-f14f-4af1-9b53-543c7cd3fae9-kube-api-access-8wchl\") pod \"ovnkube-node-xcq25\" (UID: \"958af1eb-f14f-4af1-9b53-543c7cd3fae9\") " pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.499687 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.574723 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" event={"ID":"958af1eb-f14f-4af1-9b53-543c7cd3fae9","Type":"ContainerStarted","Data":"6cb893757b94828eb92b9f512bacf89966465593f1b0a1cf24dd604709530b66"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.578011 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w862c_449d40af-fe23-4350-b66f-de4ce6614177/ovnkube-controller/3.log" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.582362 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w862c_449d40af-fe23-4350-b66f-de4ce6614177/ovn-acl-logging/0.log" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.582801 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-w862c_449d40af-fe23-4350-b66f-de4ce6614177/ovn-controller/0.log" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583095 4896 generic.go:334] "Generic (PLEG): container finished" podID="449d40af-fe23-4350-b66f-de4ce6614177" containerID="4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30" exitCode=0 Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583118 4896 generic.go:334] "Generic (PLEG): container finished" podID="449d40af-fe23-4350-b66f-de4ce6614177" containerID="d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad" exitCode=0 Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583125 4896 generic.go:334] "Generic (PLEG): container finished" podID="449d40af-fe23-4350-b66f-de4ce6614177" containerID="52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b" exitCode=0 Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583131 4896 generic.go:334] "Generic (PLEG): container finished" podID="449d40af-fe23-4350-b66f-de4ce6614177" containerID="1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc" exitCode=0 Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583138 4896 generic.go:334] "Generic (PLEG): container finished" podID="449d40af-fe23-4350-b66f-de4ce6614177" containerID="5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2" exitCode=0 Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583144 4896 generic.go:334] "Generic (PLEG): container finished" podID="449d40af-fe23-4350-b66f-de4ce6614177" containerID="3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521" exitCode=0 Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583150 4896 generic.go:334] "Generic (PLEG): container finished" podID="449d40af-fe23-4350-b66f-de4ce6614177" containerID="185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a" exitCode=143 Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583157 4896 generic.go:334] "Generic (PLEG): container finished" podID="449d40af-fe23-4350-b66f-de4ce6614177" containerID="22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf" exitCode=143 Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583190 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" event={"ID":"449d40af-fe23-4350-b66f-de4ce6614177","Type":"ContainerDied","Data":"4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583214 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" event={"ID":"449d40af-fe23-4350-b66f-de4ce6614177","Type":"ContainerDied","Data":"d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583224 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" event={"ID":"449d40af-fe23-4350-b66f-de4ce6614177","Type":"ContainerDied","Data":"52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583252 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" event={"ID":"449d40af-fe23-4350-b66f-de4ce6614177","Type":"ContainerDied","Data":"1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583266 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" event={"ID":"449d40af-fe23-4350-b66f-de4ce6614177","Type":"ContainerDied","Data":"5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583277 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" event={"ID":"449d40af-fe23-4350-b66f-de4ce6614177","Type":"ContainerDied","Data":"3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583289 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583301 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583307 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583312 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583317 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583322 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583326 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583332 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583336 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583343 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" event={"ID":"449d40af-fe23-4350-b66f-de4ce6614177","Type":"ContainerDied","Data":"185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583350 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583355 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583360 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583365 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583370 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583375 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583380 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583385 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583389 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583397 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583405 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" event={"ID":"449d40af-fe23-4350-b66f-de4ce6614177","Type":"ContainerDied","Data":"22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583412 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583419 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583424 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583430 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583434 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583439 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583444 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583449 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583453 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583458 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583464 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" event={"ID":"449d40af-fe23-4350-b66f-de4ce6614177","Type":"ContainerDied","Data":"115025435df819683983d7d2d9572a0175d2f56d4641320f8be65a2840ee3941"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583471 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583476 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583482 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583488 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583493 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583498 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583502 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583507 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583512 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583517 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583529 4896 scope.go:117] "RemoveContainer" containerID="4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.583674 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-w862c" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.591840 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fmg2f_c5ba4981-ee97-4d71-b48f-b5d12bd5911c/kube-multus/2.log" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.592294 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fmg2f_c5ba4981-ee97-4d71-b48f-b5d12bd5911c/kube-multus/1.log" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.592333 4896 generic.go:334] "Generic (PLEG): container finished" podID="c5ba4981-ee97-4d71-b48f-b5d12bd5911c" containerID="706e92310ef50b2beef6b34b867bc8cb63798490a8311ab1f91116a32f253cd1" exitCode=2 Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.592373 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fmg2f" event={"ID":"c5ba4981-ee97-4d71-b48f-b5d12bd5911c","Type":"ContainerDied","Data":"706e92310ef50b2beef6b34b867bc8cb63798490a8311ab1f91116a32f253cd1"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.592402 4896 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"54f94d214d96548b9d5a424bc162f56ffecc188391e16ee0c3b4009d55ca0472"} Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.592904 4896 scope.go:117] "RemoveContainer" containerID="706e92310ef50b2beef6b34b867bc8cb63798490a8311ab1f91116a32f253cd1" Feb 18 15:05:56 crc kubenswrapper[4896]: E0218 15:05:56.593096 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-fmg2f_openshift-multus(c5ba4981-ee97-4d71-b48f-b5d12bd5911c)\"" pod="openshift-multus/multus-fmg2f" podUID="c5ba4981-ee97-4d71-b48f-b5d12bd5911c" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.613461 4896 scope.go:117] "RemoveContainer" containerID="670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.629090 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-w862c"] Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.633656 4896 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-w862c"] Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.684223 4896 scope.go:117] "RemoveContainer" containerID="d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.696131 4896 scope.go:117] "RemoveContainer" containerID="52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.707026 4896 scope.go:117] "RemoveContainer" containerID="1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.717995 4896 scope.go:117] "RemoveContainer" containerID="5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.731708 4896 scope.go:117] "RemoveContainer" containerID="3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.742481 4896 scope.go:117] "RemoveContainer" containerID="185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.752957 4896 scope.go:117] "RemoveContainer" containerID="22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.767469 4896 scope.go:117] "RemoveContainer" containerID="aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.780823 4896 scope.go:117] "RemoveContainer" containerID="4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30" Feb 18 15:05:56 crc kubenswrapper[4896]: E0218 15:05:56.781330 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30\": container with ID starting with 4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30 not found: ID does not exist" containerID="4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.781365 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30"} err="failed to get container status \"4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30\": rpc error: code = NotFound desc = could not find container \"4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30\": container with ID starting with 4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30 not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.781388 4896 scope.go:117] "RemoveContainer" containerID="670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407" Feb 18 15:05:56 crc kubenswrapper[4896]: E0218 15:05:56.781724 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407\": container with ID starting with 670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407 not found: ID does not exist" containerID="670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.781752 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407"} err="failed to get container status \"670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407\": rpc error: code = NotFound desc = could not find container \"670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407\": container with ID starting with 670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407 not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.781768 4896 scope.go:117] "RemoveContainer" containerID="d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad" Feb 18 15:05:56 crc kubenswrapper[4896]: E0218 15:05:56.782096 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\": container with ID starting with d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad not found: ID does not exist" containerID="d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.782122 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad"} err="failed to get container status \"d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\": rpc error: code = NotFound desc = could not find container \"d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\": container with ID starting with d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.782141 4896 scope.go:117] "RemoveContainer" containerID="52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b" Feb 18 15:05:56 crc kubenswrapper[4896]: E0218 15:05:56.782862 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\": container with ID starting with 52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b not found: ID does not exist" containerID="52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.782889 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b"} err="failed to get container status \"52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\": rpc error: code = NotFound desc = could not find container \"52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\": container with ID starting with 52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.782902 4896 scope.go:117] "RemoveContainer" containerID="1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc" Feb 18 15:05:56 crc kubenswrapper[4896]: E0218 15:05:56.783172 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\": container with ID starting with 1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc not found: ID does not exist" containerID="1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.783211 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc"} err="failed to get container status \"1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\": rpc error: code = NotFound desc = could not find container \"1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\": container with ID starting with 1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.783264 4896 scope.go:117] "RemoveContainer" containerID="5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2" Feb 18 15:05:56 crc kubenswrapper[4896]: E0218 15:05:56.783560 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\": container with ID starting with 5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2 not found: ID does not exist" containerID="5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.783592 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2"} err="failed to get container status \"5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\": rpc error: code = NotFound desc = could not find container \"5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\": container with ID starting with 5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2 not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.783609 4896 scope.go:117] "RemoveContainer" containerID="3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521" Feb 18 15:05:56 crc kubenswrapper[4896]: E0218 15:05:56.783890 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\": container with ID starting with 3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521 not found: ID does not exist" containerID="3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.783943 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521"} err="failed to get container status \"3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\": rpc error: code = NotFound desc = could not find container \"3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\": container with ID starting with 3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521 not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.783961 4896 scope.go:117] "RemoveContainer" containerID="185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a" Feb 18 15:05:56 crc kubenswrapper[4896]: E0218 15:05:56.784201 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\": container with ID starting with 185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a not found: ID does not exist" containerID="185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.784230 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a"} err="failed to get container status \"185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\": rpc error: code = NotFound desc = could not find container \"185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\": container with ID starting with 185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.784265 4896 scope.go:117] "RemoveContainer" containerID="22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf" Feb 18 15:05:56 crc kubenswrapper[4896]: E0218 15:05:56.784499 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\": container with ID starting with 22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf not found: ID does not exist" containerID="22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.784522 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf"} err="failed to get container status \"22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\": rpc error: code = NotFound desc = could not find container \"22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\": container with ID starting with 22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.784541 4896 scope.go:117] "RemoveContainer" containerID="aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1" Feb 18 15:05:56 crc kubenswrapper[4896]: E0218 15:05:56.784824 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\": container with ID starting with aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1 not found: ID does not exist" containerID="aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.784850 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1"} err="failed to get container status \"aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\": rpc error: code = NotFound desc = could not find container \"aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\": container with ID starting with aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1 not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.784866 4896 scope.go:117] "RemoveContainer" containerID="4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.785150 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30"} err="failed to get container status \"4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30\": rpc error: code = NotFound desc = could not find container \"4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30\": container with ID starting with 4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30 not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.785186 4896 scope.go:117] "RemoveContainer" containerID="670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.785522 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407"} err="failed to get container status \"670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407\": rpc error: code = NotFound desc = could not find container \"670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407\": container with ID starting with 670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407 not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.785551 4896 scope.go:117] "RemoveContainer" containerID="d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.785913 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad"} err="failed to get container status \"d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\": rpc error: code = NotFound desc = could not find container \"d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\": container with ID starting with d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.785946 4896 scope.go:117] "RemoveContainer" containerID="52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.786182 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b"} err="failed to get container status \"52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\": rpc error: code = NotFound desc = could not find container \"52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\": container with ID starting with 52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.786209 4896 scope.go:117] "RemoveContainer" containerID="1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.786618 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc"} err="failed to get container status \"1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\": rpc error: code = NotFound desc = could not find container \"1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\": container with ID starting with 1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.786644 4896 scope.go:117] "RemoveContainer" containerID="5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.786892 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2"} err="failed to get container status \"5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\": rpc error: code = NotFound desc = could not find container \"5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\": container with ID starting with 5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2 not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.786917 4896 scope.go:117] "RemoveContainer" containerID="3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.787137 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521"} err="failed to get container status \"3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\": rpc error: code = NotFound desc = could not find container \"3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\": container with ID starting with 3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521 not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.787163 4896 scope.go:117] "RemoveContainer" containerID="185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.787412 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a"} err="failed to get container status \"185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\": rpc error: code = NotFound desc = could not find container \"185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\": container with ID starting with 185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.787431 4896 scope.go:117] "RemoveContainer" containerID="22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.787658 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf"} err="failed to get container status \"22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\": rpc error: code = NotFound desc = could not find container \"22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\": container with ID starting with 22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.787695 4896 scope.go:117] "RemoveContainer" containerID="aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.787950 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1"} err="failed to get container status \"aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\": rpc error: code = NotFound desc = could not find container \"aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\": container with ID starting with aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1 not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.787973 4896 scope.go:117] "RemoveContainer" containerID="4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.788164 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30"} err="failed to get container status \"4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30\": rpc error: code = NotFound desc = could not find container \"4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30\": container with ID starting with 4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30 not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.788182 4896 scope.go:117] "RemoveContainer" containerID="670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.788390 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407"} err="failed to get container status \"670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407\": rpc error: code = NotFound desc = could not find container \"670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407\": container with ID starting with 670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407 not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.788408 4896 scope.go:117] "RemoveContainer" containerID="d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.788573 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad"} err="failed to get container status \"d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\": rpc error: code = NotFound desc = could not find container \"d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\": container with ID starting with d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.788591 4896 scope.go:117] "RemoveContainer" containerID="52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.788827 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b"} err="failed to get container status \"52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\": rpc error: code = NotFound desc = could not find container \"52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\": container with ID starting with 52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.788860 4896 scope.go:117] "RemoveContainer" containerID="1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.789037 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc"} err="failed to get container status \"1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\": rpc error: code = NotFound desc = could not find container \"1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\": container with ID starting with 1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.789068 4896 scope.go:117] "RemoveContainer" containerID="5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.789339 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2"} err="failed to get container status \"5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\": rpc error: code = NotFound desc = could not find container \"5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\": container with ID starting with 5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2 not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.789359 4896 scope.go:117] "RemoveContainer" containerID="3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.789634 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521"} err="failed to get container status \"3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\": rpc error: code = NotFound desc = could not find container \"3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\": container with ID starting with 3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521 not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.789658 4896 scope.go:117] "RemoveContainer" containerID="185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.790439 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a"} err="failed to get container status \"185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\": rpc error: code = NotFound desc = could not find container \"185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\": container with ID starting with 185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.790460 4896 scope.go:117] "RemoveContainer" containerID="22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.790771 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf"} err="failed to get container status \"22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\": rpc error: code = NotFound desc = could not find container \"22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\": container with ID starting with 22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.790799 4896 scope.go:117] "RemoveContainer" containerID="aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.791059 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1"} err="failed to get container status \"aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\": rpc error: code = NotFound desc = could not find container \"aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\": container with ID starting with aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1 not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.791084 4896 scope.go:117] "RemoveContainer" containerID="4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.791262 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30"} err="failed to get container status \"4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30\": rpc error: code = NotFound desc = could not find container \"4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30\": container with ID starting with 4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30 not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.791290 4896 scope.go:117] "RemoveContainer" containerID="670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.791537 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407"} err="failed to get container status \"670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407\": rpc error: code = NotFound desc = could not find container \"670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407\": container with ID starting with 670366680e84981f402b8bb7422ecf6beaed2ed6254bb89f4e8fb56080f73407 not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.791551 4896 scope.go:117] "RemoveContainer" containerID="d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.791810 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad"} err="failed to get container status \"d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\": rpc error: code = NotFound desc = could not find container \"d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad\": container with ID starting with d4454d10f58b8e12900ade1fa34afacec811f99def4f7e3986f0247ec0e9bfad not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.791829 4896 scope.go:117] "RemoveContainer" containerID="52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.792047 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b"} err="failed to get container status \"52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\": rpc error: code = NotFound desc = could not find container \"52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b\": container with ID starting with 52c87e48243b4d2fe2c01452cb651be586ca4ae8c52b2717cd9456f18391045b not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.792071 4896 scope.go:117] "RemoveContainer" containerID="1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.792408 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc"} err="failed to get container status \"1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\": rpc error: code = NotFound desc = could not find container \"1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc\": container with ID starting with 1bd26c63ca70121d62cda83315d8e40534dd0281b48b5824dff4e19d2b44e3fc not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.792427 4896 scope.go:117] "RemoveContainer" containerID="5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.792670 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2"} err="failed to get container status \"5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\": rpc error: code = NotFound desc = could not find container \"5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2\": container with ID starting with 5847d042085624dc43ef81875773ffd24152b2fd2522ac9a04b9a5bae56ee7e2 not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.792707 4896 scope.go:117] "RemoveContainer" containerID="3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.793037 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521"} err="failed to get container status \"3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\": rpc error: code = NotFound desc = could not find container \"3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521\": container with ID starting with 3a64b2bde33e6a747254f2e4a1e69ef40e5e257462ebce331bb0384dae223521 not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.793059 4896 scope.go:117] "RemoveContainer" containerID="185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.793338 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a"} err="failed to get container status \"185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\": rpc error: code = NotFound desc = could not find container \"185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a\": container with ID starting with 185d82ac44a9f496062f7e888b3456d6e10a26f6509530138b57e678388f274a not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.793368 4896 scope.go:117] "RemoveContainer" containerID="22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.793567 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf"} err="failed to get container status \"22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\": rpc error: code = NotFound desc = could not find container \"22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf\": container with ID starting with 22e4570ba7e5f13966d152d7da3ad261d8f82b1f1ca278b0bdfce0ad2c1f22cf not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.793590 4896 scope.go:117] "RemoveContainer" containerID="aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.793857 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1"} err="failed to get container status \"aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\": rpc error: code = NotFound desc = could not find container \"aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1\": container with ID starting with aff8f4cc2591ab7ef78b0569996b651687c63fce9d95fad1f107a3cf193858b1 not found: ID does not exist" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.793879 4896 scope.go:117] "RemoveContainer" containerID="4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30" Feb 18 15:05:56 crc kubenswrapper[4896]: I0218 15:05:56.794369 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30"} err="failed to get container status \"4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30\": rpc error: code = NotFound desc = could not find container \"4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30\": container with ID starting with 4695ef9f788eba893a66df8bd65f0eccfc51b69074cd87fd852648e0bffafd30 not found: ID does not exist" Feb 18 15:05:57 crc kubenswrapper[4896]: I0218 15:05:57.599563 4896 generic.go:334] "Generic (PLEG): container finished" podID="958af1eb-f14f-4af1-9b53-543c7cd3fae9" containerID="09a2e32ded131f643c4e7d9481d29ce7e5f6868efc9c0cdafac9d1d9d148f1d0" exitCode=0 Feb 18 15:05:57 crc kubenswrapper[4896]: I0218 15:05:57.599610 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" event={"ID":"958af1eb-f14f-4af1-9b53-543c7cd3fae9","Type":"ContainerDied","Data":"09a2e32ded131f643c4e7d9481d29ce7e5f6868efc9c0cdafac9d1d9d148f1d0"} Feb 18 15:05:57 crc kubenswrapper[4896]: I0218 15:05:57.937785 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="449d40af-fe23-4350-b66f-de4ce6614177" path="/var/lib/kubelet/pods/449d40af-fe23-4350-b66f-de4ce6614177/volumes" Feb 18 15:05:58 crc kubenswrapper[4896]: I0218 15:05:58.609121 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" event={"ID":"958af1eb-f14f-4af1-9b53-543c7cd3fae9","Type":"ContainerStarted","Data":"e13a57271e64f6590ef4f04f3c67a392dc44a1e15e0cc76f22d01aab75fb473b"} Feb 18 15:05:58 crc kubenswrapper[4896]: I0218 15:05:58.609371 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" event={"ID":"958af1eb-f14f-4af1-9b53-543c7cd3fae9","Type":"ContainerStarted","Data":"58ef8768f0f164f99b0a392d707820b7d3228603a39f560087f4bc03b512f003"} Feb 18 15:05:58 crc kubenswrapper[4896]: I0218 15:05:58.609383 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" event={"ID":"958af1eb-f14f-4af1-9b53-543c7cd3fae9","Type":"ContainerStarted","Data":"e7d03b2e142be54f354a1801a51cd4c1c5001bace99551cac99836d3670077f4"} Feb 18 15:05:58 crc kubenswrapper[4896]: I0218 15:05:58.609391 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" event={"ID":"958af1eb-f14f-4af1-9b53-543c7cd3fae9","Type":"ContainerStarted","Data":"bce9579612e8713684a379d9168f3c4c7914d3ab0a5539b3c5a1463f84448d55"} Feb 18 15:05:58 crc kubenswrapper[4896]: I0218 15:05:58.609401 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" event={"ID":"958af1eb-f14f-4af1-9b53-543c7cd3fae9","Type":"ContainerStarted","Data":"fd14154dd714904d1469e968b1cc6f0401721d728d522b682f48f558aa91e7a1"} Feb 18 15:05:58 crc kubenswrapper[4896]: I0218 15:05:58.609409 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" event={"ID":"958af1eb-f14f-4af1-9b53-543c7cd3fae9","Type":"ContainerStarted","Data":"d5b33a5cf7a6a754f41d4ed75a07886fd879757f57b9cfeea4bb70ff1edfbeaa"} Feb 18 15:06:00 crc kubenswrapper[4896]: I0218 15:06:00.623003 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" event={"ID":"958af1eb-f14f-4af1-9b53-543c7cd3fae9","Type":"ContainerStarted","Data":"e27772eb0dbb43297369e88384361f668c8363455a33fde7cf24e6085d7b4672"} Feb 18 15:06:02 crc kubenswrapper[4896]: I0218 15:06:02.138651 4896 scope.go:117] "RemoveContainer" containerID="54f94d214d96548b9d5a424bc162f56ffecc188391e16ee0c3b4009d55ca0472" Feb 18 15:06:02 crc kubenswrapper[4896]: I0218 15:06:02.633873 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fmg2f_c5ba4981-ee97-4d71-b48f-b5d12bd5911c/kube-multus/2.log" Feb 18 15:06:03 crc kubenswrapper[4896]: I0218 15:06:03.641661 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" event={"ID":"958af1eb-f14f-4af1-9b53-543c7cd3fae9","Type":"ContainerStarted","Data":"2099627da36288e0917d5c1b20ed05c83ce9edcb079c9e08d031fdc1af9da344"} Feb 18 15:06:03 crc kubenswrapper[4896]: I0218 15:06:03.642848 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:06:03 crc kubenswrapper[4896]: I0218 15:06:03.642881 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:06:03 crc kubenswrapper[4896]: I0218 15:06:03.642891 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:06:03 crc kubenswrapper[4896]: I0218 15:06:03.667982 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:06:03 crc kubenswrapper[4896]: I0218 15:06:03.668324 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:06:03 crc kubenswrapper[4896]: I0218 15:06:03.674088 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" podStartSLOduration=7.674074367 podStartE2EDuration="7.674074367s" podCreationTimestamp="2026-02-18 15:05:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:06:03.673363004 +0000 UTC m=+662.602697217" watchObservedRunningTime="2026-02-18 15:06:03.674074367 +0000 UTC m=+662.603408580" Feb 18 15:06:10 crc kubenswrapper[4896]: I0218 15:06:10.927926 4896 scope.go:117] "RemoveContainer" containerID="706e92310ef50b2beef6b34b867bc8cb63798490a8311ab1f91116a32f253cd1" Feb 18 15:06:10 crc kubenswrapper[4896]: E0218 15:06:10.928624 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-fmg2f_openshift-multus(c5ba4981-ee97-4d71-b48f-b5d12bd5911c)\"" pod="openshift-multus/multus-fmg2f" podUID="c5ba4981-ee97-4d71-b48f-b5d12bd5911c" Feb 18 15:06:16 crc kubenswrapper[4896]: I0218 15:06:16.993009 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7"] Feb 18 15:06:16 crc kubenswrapper[4896]: I0218 15:06:16.994713 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7" Feb 18 15:06:16 crc kubenswrapper[4896]: I0218 15:06:16.996681 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 18 15:06:17 crc kubenswrapper[4896]: I0218 15:06:17.006139 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7"] Feb 18 15:06:17 crc kubenswrapper[4896]: I0218 15:06:17.089173 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7\" (UID: \"314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7" Feb 18 15:06:17 crc kubenswrapper[4896]: I0218 15:06:17.089511 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ljlt\" (UniqueName: \"kubernetes.io/projected/314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1-kube-api-access-5ljlt\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7\" (UID: \"314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7" Feb 18 15:06:17 crc kubenswrapper[4896]: I0218 15:06:17.089805 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7\" (UID: \"314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7" Feb 18 15:06:17 crc kubenswrapper[4896]: I0218 15:06:17.190599 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ljlt\" (UniqueName: \"kubernetes.io/projected/314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1-kube-api-access-5ljlt\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7\" (UID: \"314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7" Feb 18 15:06:17 crc kubenswrapper[4896]: I0218 15:06:17.190687 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7\" (UID: \"314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7" Feb 18 15:06:17 crc kubenswrapper[4896]: I0218 15:06:17.190741 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7\" (UID: \"314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7" Feb 18 15:06:17 crc kubenswrapper[4896]: I0218 15:06:17.191365 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1-util\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7\" (UID: \"314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7" Feb 18 15:06:17 crc kubenswrapper[4896]: I0218 15:06:17.191618 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1-bundle\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7\" (UID: \"314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7" Feb 18 15:06:17 crc kubenswrapper[4896]: I0218 15:06:17.211642 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ljlt\" (UniqueName: \"kubernetes.io/projected/314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1-kube-api-access-5ljlt\") pod \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7\" (UID: \"314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1\") " pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7" Feb 18 15:06:17 crc kubenswrapper[4896]: I0218 15:06:17.370423 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7" Feb 18 15:06:17 crc kubenswrapper[4896]: E0218 15:06:17.407460 4896 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7_openshift-marketplace_314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1_0(891ed69f7e54404d2a44cd6d7f24ebb6b8a022d8e2dd39270e93e3b52b086f3f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 18 15:06:17 crc kubenswrapper[4896]: E0218 15:06:17.407563 4896 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7_openshift-marketplace_314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1_0(891ed69f7e54404d2a44cd6d7f24ebb6b8a022d8e2dd39270e93e3b52b086f3f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7" Feb 18 15:06:17 crc kubenswrapper[4896]: E0218 15:06:17.407612 4896 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7_openshift-marketplace_314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1_0(891ed69f7e54404d2a44cd6d7f24ebb6b8a022d8e2dd39270e93e3b52b086f3f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7" Feb 18 15:06:17 crc kubenswrapper[4896]: E0218 15:06:17.407702 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7_openshift-marketplace(314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7_openshift-marketplace(314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7_openshift-marketplace_314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1_0(891ed69f7e54404d2a44cd6d7f24ebb6b8a022d8e2dd39270e93e3b52b086f3f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7" podUID="314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1" Feb 18 15:06:17 crc kubenswrapper[4896]: I0218 15:06:17.707398 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7" Feb 18 15:06:17 crc kubenswrapper[4896]: I0218 15:06:17.707986 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7" Feb 18 15:06:17 crc kubenswrapper[4896]: E0218 15:06:17.742508 4896 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7_openshift-marketplace_314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1_0(d781817d866fdd1fb91c34f0c030dadd95bcac2afd4ad6cafa1232727876b059): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 18 15:06:17 crc kubenswrapper[4896]: E0218 15:06:17.742604 4896 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7_openshift-marketplace_314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1_0(d781817d866fdd1fb91c34f0c030dadd95bcac2afd4ad6cafa1232727876b059): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7" Feb 18 15:06:17 crc kubenswrapper[4896]: E0218 15:06:17.742642 4896 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7_openshift-marketplace_314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1_0(d781817d866fdd1fb91c34f0c030dadd95bcac2afd4ad6cafa1232727876b059): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7" Feb 18 15:06:17 crc kubenswrapper[4896]: E0218 15:06:17.742718 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7_openshift-marketplace(314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7_openshift-marketplace(314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7_openshift-marketplace_314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1_0(d781817d866fdd1fb91c34f0c030dadd95bcac2afd4ad6cafa1232727876b059): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7" podUID="314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1" Feb 18 15:06:25 crc kubenswrapper[4896]: I0218 15:06:25.928282 4896 scope.go:117] "RemoveContainer" containerID="706e92310ef50b2beef6b34b867bc8cb63798490a8311ab1f91116a32f253cd1" Feb 18 15:06:26 crc kubenswrapper[4896]: I0218 15:06:26.526352 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-xcq25" Feb 18 15:06:26 crc kubenswrapper[4896]: I0218 15:06:26.758888 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-fmg2f_c5ba4981-ee97-4d71-b48f-b5d12bd5911c/kube-multus/2.log" Feb 18 15:06:26 crc kubenswrapper[4896]: I0218 15:06:26.758953 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-fmg2f" event={"ID":"c5ba4981-ee97-4d71-b48f-b5d12bd5911c","Type":"ContainerStarted","Data":"37ead64ba8a01411aee9ef33b2b76a8d2877e6342848f16c858fb83c2a7cb884"} Feb 18 15:06:28 crc kubenswrapper[4896]: I0218 15:06:28.927414 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7" Feb 18 15:06:28 crc kubenswrapper[4896]: I0218 15:06:28.928191 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7" Feb 18 15:06:29 crc kubenswrapper[4896]: I0218 15:06:29.099276 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7"] Feb 18 15:06:29 crc kubenswrapper[4896]: I0218 15:06:29.774896 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7" event={"ID":"314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1","Type":"ContainerStarted","Data":"c776d17960d683672c1bb4f77d07d9a3397b630cb19731f776cb840f8c989f97"} Feb 18 15:06:29 crc kubenswrapper[4896]: I0218 15:06:29.775511 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7" event={"ID":"314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1","Type":"ContainerStarted","Data":"b048202c1d7ea4854ab4cedb18fd50fd267245bd8d92f390b2f3526052db8495"} Feb 18 15:06:30 crc kubenswrapper[4896]: I0218 15:06:30.780208 4896 generic.go:334] "Generic (PLEG): container finished" podID="314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1" containerID="c776d17960d683672c1bb4f77d07d9a3397b630cb19731f776cb840f8c989f97" exitCode=0 Feb 18 15:06:30 crc kubenswrapper[4896]: I0218 15:06:30.780268 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7" event={"ID":"314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1","Type":"ContainerDied","Data":"c776d17960d683672c1bb4f77d07d9a3397b630cb19731f776cb840f8c989f97"} Feb 18 15:06:30 crc kubenswrapper[4896]: I0218 15:06:30.782608 4896 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 18 15:06:32 crc kubenswrapper[4896]: I0218 15:06:32.793922 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7" event={"ID":"314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1","Type":"ContainerStarted","Data":"07a1ffed7378db2e18cfaec1c46215c823c7c9f9df07623b13dfe7d97007c9a1"} Feb 18 15:06:33 crc kubenswrapper[4896]: I0218 15:06:33.803889 4896 generic.go:334] "Generic (PLEG): container finished" podID="314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1" containerID="07a1ffed7378db2e18cfaec1c46215c823c7c9f9df07623b13dfe7d97007c9a1" exitCode=0 Feb 18 15:06:33 crc kubenswrapper[4896]: I0218 15:06:33.804011 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7" event={"ID":"314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1","Type":"ContainerDied","Data":"07a1ffed7378db2e18cfaec1c46215c823c7c9f9df07623b13dfe7d97007c9a1"} Feb 18 15:06:34 crc kubenswrapper[4896]: I0218 15:06:34.814013 4896 generic.go:334] "Generic (PLEG): container finished" podID="314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1" containerID="88076fbe07c2c4d9f6c90587479005c2dde200973480de17070c38d084b335fe" exitCode=0 Feb 18 15:06:34 crc kubenswrapper[4896]: I0218 15:06:34.814086 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7" event={"ID":"314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1","Type":"ContainerDied","Data":"88076fbe07c2c4d9f6c90587479005c2dde200973480de17070c38d084b335fe"} Feb 18 15:06:36 crc kubenswrapper[4896]: I0218 15:06:36.047553 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7" Feb 18 15:06:36 crc kubenswrapper[4896]: I0218 15:06:36.146212 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ljlt\" (UniqueName: \"kubernetes.io/projected/314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1-kube-api-access-5ljlt\") pod \"314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1\" (UID: \"314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1\") " Feb 18 15:06:36 crc kubenswrapper[4896]: I0218 15:06:36.146306 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1-util\") pod \"314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1\" (UID: \"314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1\") " Feb 18 15:06:36 crc kubenswrapper[4896]: I0218 15:06:36.146328 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1-bundle\") pod \"314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1\" (UID: \"314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1\") " Feb 18 15:06:36 crc kubenswrapper[4896]: I0218 15:06:36.147357 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1-bundle" (OuterVolumeSpecName: "bundle") pod "314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1" (UID: "314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:06:36 crc kubenswrapper[4896]: I0218 15:06:36.152403 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1-kube-api-access-5ljlt" (OuterVolumeSpecName: "kube-api-access-5ljlt") pod "314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1" (UID: "314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1"). InnerVolumeSpecName "kube-api-access-5ljlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:06:36 crc kubenswrapper[4896]: I0218 15:06:36.158485 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1-util" (OuterVolumeSpecName: "util") pod "314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1" (UID: "314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:06:36 crc kubenswrapper[4896]: I0218 15:06:36.247209 4896 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1-util\") on node \"crc\" DevicePath \"\"" Feb 18 15:06:36 crc kubenswrapper[4896]: I0218 15:06:36.247278 4896 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 15:06:36 crc kubenswrapper[4896]: I0218 15:06:36.247287 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ljlt\" (UniqueName: \"kubernetes.io/projected/314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1-kube-api-access-5ljlt\") on node \"crc\" DevicePath \"\"" Feb 18 15:06:36 crc kubenswrapper[4896]: I0218 15:06:36.826120 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7" event={"ID":"314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1","Type":"ContainerDied","Data":"b048202c1d7ea4854ab4cedb18fd50fd267245bd8d92f390b2f3526052db8495"} Feb 18 15:06:36 crc kubenswrapper[4896]: I0218 15:06:36.826163 4896 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b048202c1d7ea4854ab4cedb18fd50fd267245bd8d92f390b2f3526052db8495" Feb 18 15:06:36 crc kubenswrapper[4896]: I0218 15:06:36.826172 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.177358 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-756bb4b59-vd48g"] Feb 18 15:06:45 crc kubenswrapper[4896]: E0218 15:06:45.178144 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1" containerName="util" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.178159 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1" containerName="util" Feb 18 15:06:45 crc kubenswrapper[4896]: E0218 15:06:45.178169 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1" containerName="extract" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.178177 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1" containerName="extract" Feb 18 15:06:45 crc kubenswrapper[4896]: E0218 15:06:45.178186 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1" containerName="pull" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.178193 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1" containerName="pull" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.178341 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1" containerName="extract" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.178771 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-756bb4b59-vd48g" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.187440 4896 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.187853 4896 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-sgdss" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.187879 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.188213 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.193023 4896 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.199264 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-756bb4b59-vd48g"] Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.253869 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/60827093-1be2-4ca9-b9d3-3caaf37ebe83-apiservice-cert\") pod \"metallb-operator-controller-manager-756bb4b59-vd48g\" (UID: \"60827093-1be2-4ca9-b9d3-3caaf37ebe83\") " pod="metallb-system/metallb-operator-controller-manager-756bb4b59-vd48g" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.253924 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/60827093-1be2-4ca9-b9d3-3caaf37ebe83-webhook-cert\") pod \"metallb-operator-controller-manager-756bb4b59-vd48g\" (UID: \"60827093-1be2-4ca9-b9d3-3caaf37ebe83\") " pod="metallb-system/metallb-operator-controller-manager-756bb4b59-vd48g" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.253967 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cf2b8\" (UniqueName: \"kubernetes.io/projected/60827093-1be2-4ca9-b9d3-3caaf37ebe83-kube-api-access-cf2b8\") pod \"metallb-operator-controller-manager-756bb4b59-vd48g\" (UID: \"60827093-1be2-4ca9-b9d3-3caaf37ebe83\") " pod="metallb-system/metallb-operator-controller-manager-756bb4b59-vd48g" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.355758 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/60827093-1be2-4ca9-b9d3-3caaf37ebe83-apiservice-cert\") pod \"metallb-operator-controller-manager-756bb4b59-vd48g\" (UID: \"60827093-1be2-4ca9-b9d3-3caaf37ebe83\") " pod="metallb-system/metallb-operator-controller-manager-756bb4b59-vd48g" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.355875 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/60827093-1be2-4ca9-b9d3-3caaf37ebe83-webhook-cert\") pod \"metallb-operator-controller-manager-756bb4b59-vd48g\" (UID: \"60827093-1be2-4ca9-b9d3-3caaf37ebe83\") " pod="metallb-system/metallb-operator-controller-manager-756bb4b59-vd48g" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.355917 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cf2b8\" (UniqueName: \"kubernetes.io/projected/60827093-1be2-4ca9-b9d3-3caaf37ebe83-kube-api-access-cf2b8\") pod \"metallb-operator-controller-manager-756bb4b59-vd48g\" (UID: \"60827093-1be2-4ca9-b9d3-3caaf37ebe83\") " pod="metallb-system/metallb-operator-controller-manager-756bb4b59-vd48g" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.368955 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/60827093-1be2-4ca9-b9d3-3caaf37ebe83-webhook-cert\") pod \"metallb-operator-controller-manager-756bb4b59-vd48g\" (UID: \"60827093-1be2-4ca9-b9d3-3caaf37ebe83\") " pod="metallb-system/metallb-operator-controller-manager-756bb4b59-vd48g" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.368955 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/60827093-1be2-4ca9-b9d3-3caaf37ebe83-apiservice-cert\") pod \"metallb-operator-controller-manager-756bb4b59-vd48g\" (UID: \"60827093-1be2-4ca9-b9d3-3caaf37ebe83\") " pod="metallb-system/metallb-operator-controller-manager-756bb4b59-vd48g" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.373354 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cf2b8\" (UniqueName: \"kubernetes.io/projected/60827093-1be2-4ca9-b9d3-3caaf37ebe83-kube-api-access-cf2b8\") pod \"metallb-operator-controller-manager-756bb4b59-vd48g\" (UID: \"60827093-1be2-4ca9-b9d3-3caaf37ebe83\") " pod="metallb-system/metallb-operator-controller-manager-756bb4b59-vd48g" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.407204 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-d66b4dc77-fqp6g"] Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.407787 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-d66b4dc77-fqp6g" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.409770 4896 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-c4qbq" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.409834 4896 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.410109 4896 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.453645 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-d66b4dc77-fqp6g"] Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.457437 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5bp9\" (UniqueName: \"kubernetes.io/projected/23fd11e5-8c0f-434f-a334-01bf16e764e8-kube-api-access-j5bp9\") pod \"metallb-operator-webhook-server-d66b4dc77-fqp6g\" (UID: \"23fd11e5-8c0f-434f-a334-01bf16e764e8\") " pod="metallb-system/metallb-operator-webhook-server-d66b4dc77-fqp6g" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.457493 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/23fd11e5-8c0f-434f-a334-01bf16e764e8-webhook-cert\") pod \"metallb-operator-webhook-server-d66b4dc77-fqp6g\" (UID: \"23fd11e5-8c0f-434f-a334-01bf16e764e8\") " pod="metallb-system/metallb-operator-webhook-server-d66b4dc77-fqp6g" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.457528 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/23fd11e5-8c0f-434f-a334-01bf16e764e8-apiservice-cert\") pod \"metallb-operator-webhook-server-d66b4dc77-fqp6g\" (UID: \"23fd11e5-8c0f-434f-a334-01bf16e764e8\") " pod="metallb-system/metallb-operator-webhook-server-d66b4dc77-fqp6g" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.493719 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-756bb4b59-vd48g" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.559457 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/23fd11e5-8c0f-434f-a334-01bf16e764e8-apiservice-cert\") pod \"metallb-operator-webhook-server-d66b4dc77-fqp6g\" (UID: \"23fd11e5-8c0f-434f-a334-01bf16e764e8\") " pod="metallb-system/metallb-operator-webhook-server-d66b4dc77-fqp6g" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.559592 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5bp9\" (UniqueName: \"kubernetes.io/projected/23fd11e5-8c0f-434f-a334-01bf16e764e8-kube-api-access-j5bp9\") pod \"metallb-operator-webhook-server-d66b4dc77-fqp6g\" (UID: \"23fd11e5-8c0f-434f-a334-01bf16e764e8\") " pod="metallb-system/metallb-operator-webhook-server-d66b4dc77-fqp6g" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.559636 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/23fd11e5-8c0f-434f-a334-01bf16e764e8-webhook-cert\") pod \"metallb-operator-webhook-server-d66b4dc77-fqp6g\" (UID: \"23fd11e5-8c0f-434f-a334-01bf16e764e8\") " pod="metallb-system/metallb-operator-webhook-server-d66b4dc77-fqp6g" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.563457 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/23fd11e5-8c0f-434f-a334-01bf16e764e8-webhook-cert\") pod \"metallb-operator-webhook-server-d66b4dc77-fqp6g\" (UID: \"23fd11e5-8c0f-434f-a334-01bf16e764e8\") " pod="metallb-system/metallb-operator-webhook-server-d66b4dc77-fqp6g" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.568730 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/23fd11e5-8c0f-434f-a334-01bf16e764e8-apiservice-cert\") pod \"metallb-operator-webhook-server-d66b4dc77-fqp6g\" (UID: \"23fd11e5-8c0f-434f-a334-01bf16e764e8\") " pod="metallb-system/metallb-operator-webhook-server-d66b4dc77-fqp6g" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.574424 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5bp9\" (UniqueName: \"kubernetes.io/projected/23fd11e5-8c0f-434f-a334-01bf16e764e8-kube-api-access-j5bp9\") pod \"metallb-operator-webhook-server-d66b4dc77-fqp6g\" (UID: \"23fd11e5-8c0f-434f-a334-01bf16e764e8\") " pod="metallb-system/metallb-operator-webhook-server-d66b4dc77-fqp6g" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.720144 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-d66b4dc77-fqp6g" Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.753081 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-756bb4b59-vd48g"] Feb 18 15:06:45 crc kubenswrapper[4896]: I0218 15:06:45.871648 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-756bb4b59-vd48g" event={"ID":"60827093-1be2-4ca9-b9d3-3caaf37ebe83","Type":"ContainerStarted","Data":"b58eca036593ef65848ade91377529897f3fe3f12def3a5169adcc1bd23f5999"} Feb 18 15:06:46 crc kubenswrapper[4896]: I0218 15:06:46.121361 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-d66b4dc77-fqp6g"] Feb 18 15:06:46 crc kubenswrapper[4896]: W0218 15:06:46.125642 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23fd11e5_8c0f_434f_a334_01bf16e764e8.slice/crio-20d575844953a2969e7d0be10d8fdb082c07d6fab652aa5f286a64aa8257cc8e WatchSource:0}: Error finding container 20d575844953a2969e7d0be10d8fdb082c07d6fab652aa5f286a64aa8257cc8e: Status 404 returned error can't find the container with id 20d575844953a2969e7d0be10d8fdb082c07d6fab652aa5f286a64aa8257cc8e Feb 18 15:06:46 crc kubenswrapper[4896]: I0218 15:06:46.876767 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-d66b4dc77-fqp6g" event={"ID":"23fd11e5-8c0f-434f-a334-01bf16e764e8","Type":"ContainerStarted","Data":"20d575844953a2969e7d0be10d8fdb082c07d6fab652aa5f286a64aa8257cc8e"} Feb 18 15:06:48 crc kubenswrapper[4896]: I0218 15:06:48.889754 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-756bb4b59-vd48g" event={"ID":"60827093-1be2-4ca9-b9d3-3caaf37ebe83","Type":"ContainerStarted","Data":"cef3b2784de082e723e6e2cdde31efd5935c7637a30d692f9cf8e34046b37bc7"} Feb 18 15:06:48 crc kubenswrapper[4896]: I0218 15:06:48.889902 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-756bb4b59-vd48g" Feb 18 15:06:48 crc kubenswrapper[4896]: I0218 15:06:48.917076 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-756bb4b59-vd48g" podStartSLOduration=1.014860813 podStartE2EDuration="3.917060021s" podCreationTimestamp="2026-02-18 15:06:45 +0000 UTC" firstStartedPulling="2026-02-18 15:06:45.767504127 +0000 UTC m=+704.696838340" lastFinishedPulling="2026-02-18 15:06:48.669703335 +0000 UTC m=+707.599037548" observedRunningTime="2026-02-18 15:06:48.911265905 +0000 UTC m=+707.840600118" watchObservedRunningTime="2026-02-18 15:06:48.917060021 +0000 UTC m=+707.846394234" Feb 18 15:06:50 crc kubenswrapper[4896]: I0218 15:06:50.902820 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-d66b4dc77-fqp6g" event={"ID":"23fd11e5-8c0f-434f-a334-01bf16e764e8","Type":"ContainerStarted","Data":"cfebcae929f3b50a3b47a6daa0929315ddfb04b0e4f3080aab05063d067d2b8b"} Feb 18 15:06:50 crc kubenswrapper[4896]: I0218 15:06:50.903293 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-d66b4dc77-fqp6g" Feb 18 15:06:50 crc kubenswrapper[4896]: I0218 15:06:50.936564 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-d66b4dc77-fqp6g" podStartSLOduration=1.7323937790000001 podStartE2EDuration="5.9365479s" podCreationTimestamp="2026-02-18 15:06:45 +0000 UTC" firstStartedPulling="2026-02-18 15:06:46.128320245 +0000 UTC m=+705.057654458" lastFinishedPulling="2026-02-18 15:06:50.332474366 +0000 UTC m=+709.261808579" observedRunningTime="2026-02-18 15:06:50.925342418 +0000 UTC m=+709.854676641" watchObservedRunningTime="2026-02-18 15:06:50.9365479 +0000 UTC m=+709.865882113" Feb 18 15:07:04 crc kubenswrapper[4896]: I0218 15:07:04.076047 4896 patch_prober.go:28] interesting pod/machine-config-daemon-24rq5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:07:04 crc kubenswrapper[4896]: I0218 15:07:04.076659 4896 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:07:05 crc kubenswrapper[4896]: I0218 15:07:05.725436 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-d66b4dc77-fqp6g" Feb 18 15:07:25 crc kubenswrapper[4896]: I0218 15:07:25.496579 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-756bb4b59-vd48g" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.135277 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-jxgkk"] Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.147863 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-jxgkk" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.150435 4896 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-6vrdx" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.150836 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-lnrk2"] Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.154743 4896 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.157194 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-lnrk2" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.161493 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-jxgkk"] Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.162752 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.163016 4896 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.195899 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/29e01ca8-ce78-4769-a5d5-453834a27bb3-reloader\") pod \"frr-k8s-lnrk2\" (UID: \"29e01ca8-ce78-4769-a5d5-453834a27bb3\") " pod="metallb-system/frr-k8s-lnrk2" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.195942 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/29e01ca8-ce78-4769-a5d5-453834a27bb3-frr-startup\") pod \"frr-k8s-lnrk2\" (UID: \"29e01ca8-ce78-4769-a5d5-453834a27bb3\") " pod="metallb-system/frr-k8s-lnrk2" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.195967 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9vhb\" (UniqueName: \"kubernetes.io/projected/29e01ca8-ce78-4769-a5d5-453834a27bb3-kube-api-access-z9vhb\") pod \"frr-k8s-lnrk2\" (UID: \"29e01ca8-ce78-4769-a5d5-453834a27bb3\") " pod="metallb-system/frr-k8s-lnrk2" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.195987 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/29e01ca8-ce78-4769-a5d5-453834a27bb3-metrics-certs\") pod \"frr-k8s-lnrk2\" (UID: \"29e01ca8-ce78-4769-a5d5-453834a27bb3\") " pod="metallb-system/frr-k8s-lnrk2" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.196008 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2a8e7f2-bf46-4f7d-a7e0-c1373f5c2136-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-jxgkk\" (UID: \"c2a8e7f2-bf46-4f7d-a7e0-c1373f5c2136\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-jxgkk" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.196023 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/29e01ca8-ce78-4769-a5d5-453834a27bb3-frr-conf\") pod \"frr-k8s-lnrk2\" (UID: \"29e01ca8-ce78-4769-a5d5-453834a27bb3\") " pod="metallb-system/frr-k8s-lnrk2" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.196048 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/29e01ca8-ce78-4769-a5d5-453834a27bb3-frr-sockets\") pod \"frr-k8s-lnrk2\" (UID: \"29e01ca8-ce78-4769-a5d5-453834a27bb3\") " pod="metallb-system/frr-k8s-lnrk2" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.196064 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/29e01ca8-ce78-4769-a5d5-453834a27bb3-metrics\") pod \"frr-k8s-lnrk2\" (UID: \"29e01ca8-ce78-4769-a5d5-453834a27bb3\") " pod="metallb-system/frr-k8s-lnrk2" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.196087 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjtjj\" (UniqueName: \"kubernetes.io/projected/c2a8e7f2-bf46-4f7d-a7e0-c1373f5c2136-kube-api-access-pjtjj\") pod \"frr-k8s-webhook-server-78b44bf5bb-jxgkk\" (UID: \"c2a8e7f2-bf46-4f7d-a7e0-c1373f5c2136\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-jxgkk" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.218843 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-qb82t"] Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.219973 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-qb82t" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.221720 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-69bbfbf88f-zqgw4"] Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.222104 4896 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.222331 4896 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-5kt8z" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.222559 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-69bbfbf88f-zqgw4" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.223513 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.223721 4896 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.226185 4896 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.233855 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-69bbfbf88f-zqgw4"] Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.297678 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d2436da1-6ba8-4028-9012-1160a473af76-metallb-excludel2\") pod \"speaker-qb82t\" (UID: \"d2436da1-6ba8-4028-9012-1160a473af76\") " pod="metallb-system/speaker-qb82t" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.297725 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7d817dfd-f711-48b3-8630-83d4b1a784de-cert\") pod \"controller-69bbfbf88f-zqgw4\" (UID: \"7d817dfd-f711-48b3-8630-83d4b1a784de\") " pod="metallb-system/controller-69bbfbf88f-zqgw4" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.297765 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlgx6\" (UniqueName: \"kubernetes.io/projected/7d817dfd-f711-48b3-8630-83d4b1a784de-kube-api-access-qlgx6\") pod \"controller-69bbfbf88f-zqgw4\" (UID: \"7d817dfd-f711-48b3-8630-83d4b1a784de\") " pod="metallb-system/controller-69bbfbf88f-zqgw4" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.297792 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/29e01ca8-ce78-4769-a5d5-453834a27bb3-reloader\") pod \"frr-k8s-lnrk2\" (UID: \"29e01ca8-ce78-4769-a5d5-453834a27bb3\") " pod="metallb-system/frr-k8s-lnrk2" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.297810 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7d817dfd-f711-48b3-8630-83d4b1a784de-metrics-certs\") pod \"controller-69bbfbf88f-zqgw4\" (UID: \"7d817dfd-f711-48b3-8630-83d4b1a784de\") " pod="metallb-system/controller-69bbfbf88f-zqgw4" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.297915 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/29e01ca8-ce78-4769-a5d5-453834a27bb3-frr-startup\") pod \"frr-k8s-lnrk2\" (UID: \"29e01ca8-ce78-4769-a5d5-453834a27bb3\") " pod="metallb-system/frr-k8s-lnrk2" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.298008 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9vhb\" (UniqueName: \"kubernetes.io/projected/29e01ca8-ce78-4769-a5d5-453834a27bb3-kube-api-access-z9vhb\") pod \"frr-k8s-lnrk2\" (UID: \"29e01ca8-ce78-4769-a5d5-453834a27bb3\") " pod="metallb-system/frr-k8s-lnrk2" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.298063 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/29e01ca8-ce78-4769-a5d5-453834a27bb3-metrics-certs\") pod \"frr-k8s-lnrk2\" (UID: \"29e01ca8-ce78-4769-a5d5-453834a27bb3\") " pod="metallb-system/frr-k8s-lnrk2" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.298116 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2a8e7f2-bf46-4f7d-a7e0-c1373f5c2136-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-jxgkk\" (UID: \"c2a8e7f2-bf46-4f7d-a7e0-c1373f5c2136\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-jxgkk" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.298146 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9r2fd\" (UniqueName: \"kubernetes.io/projected/d2436da1-6ba8-4028-9012-1160a473af76-kube-api-access-9r2fd\") pod \"speaker-qb82t\" (UID: \"d2436da1-6ba8-4028-9012-1160a473af76\") " pod="metallb-system/speaker-qb82t" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.298172 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/29e01ca8-ce78-4769-a5d5-453834a27bb3-frr-conf\") pod \"frr-k8s-lnrk2\" (UID: \"29e01ca8-ce78-4769-a5d5-453834a27bb3\") " pod="metallb-system/frr-k8s-lnrk2" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.298257 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/29e01ca8-ce78-4769-a5d5-453834a27bb3-frr-sockets\") pod \"frr-k8s-lnrk2\" (UID: \"29e01ca8-ce78-4769-a5d5-453834a27bb3\") " pod="metallb-system/frr-k8s-lnrk2" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.298277 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/29e01ca8-ce78-4769-a5d5-453834a27bb3-metrics\") pod \"frr-k8s-lnrk2\" (UID: \"29e01ca8-ce78-4769-a5d5-453834a27bb3\") " pod="metallb-system/frr-k8s-lnrk2" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.298293 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/29e01ca8-ce78-4769-a5d5-453834a27bb3-reloader\") pod \"frr-k8s-lnrk2\" (UID: \"29e01ca8-ce78-4769-a5d5-453834a27bb3\") " pod="metallb-system/frr-k8s-lnrk2" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.298340 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjtjj\" (UniqueName: \"kubernetes.io/projected/c2a8e7f2-bf46-4f7d-a7e0-c1373f5c2136-kube-api-access-pjtjj\") pod \"frr-k8s-webhook-server-78b44bf5bb-jxgkk\" (UID: \"c2a8e7f2-bf46-4f7d-a7e0-c1373f5c2136\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-jxgkk" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.298376 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d2436da1-6ba8-4028-9012-1160a473af76-metrics-certs\") pod \"speaker-qb82t\" (UID: \"d2436da1-6ba8-4028-9012-1160a473af76\") " pod="metallb-system/speaker-qb82t" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.298421 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d2436da1-6ba8-4028-9012-1160a473af76-memberlist\") pod \"speaker-qb82t\" (UID: \"d2436da1-6ba8-4028-9012-1160a473af76\") " pod="metallb-system/speaker-qb82t" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.298531 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/29e01ca8-ce78-4769-a5d5-453834a27bb3-frr-conf\") pod \"frr-k8s-lnrk2\" (UID: \"29e01ca8-ce78-4769-a5d5-453834a27bb3\") " pod="metallb-system/frr-k8s-lnrk2" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.298619 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/29e01ca8-ce78-4769-a5d5-453834a27bb3-frr-sockets\") pod \"frr-k8s-lnrk2\" (UID: \"29e01ca8-ce78-4769-a5d5-453834a27bb3\") " pod="metallb-system/frr-k8s-lnrk2" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.298876 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/29e01ca8-ce78-4769-a5d5-453834a27bb3-frr-startup\") pod \"frr-k8s-lnrk2\" (UID: \"29e01ca8-ce78-4769-a5d5-453834a27bb3\") " pod="metallb-system/frr-k8s-lnrk2" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.298904 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/29e01ca8-ce78-4769-a5d5-453834a27bb3-metrics\") pod \"frr-k8s-lnrk2\" (UID: \"29e01ca8-ce78-4769-a5d5-453834a27bb3\") " pod="metallb-system/frr-k8s-lnrk2" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.303192 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/29e01ca8-ce78-4769-a5d5-453834a27bb3-metrics-certs\") pod \"frr-k8s-lnrk2\" (UID: \"29e01ca8-ce78-4769-a5d5-453834a27bb3\") " pod="metallb-system/frr-k8s-lnrk2" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.310905 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c2a8e7f2-bf46-4f7d-a7e0-c1373f5c2136-cert\") pod \"frr-k8s-webhook-server-78b44bf5bb-jxgkk\" (UID: \"c2a8e7f2-bf46-4f7d-a7e0-c1373f5c2136\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-jxgkk" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.315846 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9vhb\" (UniqueName: \"kubernetes.io/projected/29e01ca8-ce78-4769-a5d5-453834a27bb3-kube-api-access-z9vhb\") pod \"frr-k8s-lnrk2\" (UID: \"29e01ca8-ce78-4769-a5d5-453834a27bb3\") " pod="metallb-system/frr-k8s-lnrk2" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.318956 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjtjj\" (UniqueName: \"kubernetes.io/projected/c2a8e7f2-bf46-4f7d-a7e0-c1373f5c2136-kube-api-access-pjtjj\") pod \"frr-k8s-webhook-server-78b44bf5bb-jxgkk\" (UID: \"c2a8e7f2-bf46-4f7d-a7e0-c1373f5c2136\") " pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-jxgkk" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.399753 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d2436da1-6ba8-4028-9012-1160a473af76-metrics-certs\") pod \"speaker-qb82t\" (UID: \"d2436da1-6ba8-4028-9012-1160a473af76\") " pod="metallb-system/speaker-qb82t" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.399795 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d2436da1-6ba8-4028-9012-1160a473af76-memberlist\") pod \"speaker-qb82t\" (UID: \"d2436da1-6ba8-4028-9012-1160a473af76\") " pod="metallb-system/speaker-qb82t" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.399845 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d2436da1-6ba8-4028-9012-1160a473af76-metallb-excludel2\") pod \"speaker-qb82t\" (UID: \"d2436da1-6ba8-4028-9012-1160a473af76\") " pod="metallb-system/speaker-qb82t" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.399881 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7d817dfd-f711-48b3-8630-83d4b1a784de-cert\") pod \"controller-69bbfbf88f-zqgw4\" (UID: \"7d817dfd-f711-48b3-8630-83d4b1a784de\") " pod="metallb-system/controller-69bbfbf88f-zqgw4" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.399924 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlgx6\" (UniqueName: \"kubernetes.io/projected/7d817dfd-f711-48b3-8630-83d4b1a784de-kube-api-access-qlgx6\") pod \"controller-69bbfbf88f-zqgw4\" (UID: \"7d817dfd-f711-48b3-8630-83d4b1a784de\") " pod="metallb-system/controller-69bbfbf88f-zqgw4" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.399941 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7d817dfd-f711-48b3-8630-83d4b1a784de-metrics-certs\") pod \"controller-69bbfbf88f-zqgw4\" (UID: \"7d817dfd-f711-48b3-8630-83d4b1a784de\") " pod="metallb-system/controller-69bbfbf88f-zqgw4" Feb 18 15:07:26 crc kubenswrapper[4896]: E0218 15:07:26.399948 4896 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.399971 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9r2fd\" (UniqueName: \"kubernetes.io/projected/d2436da1-6ba8-4028-9012-1160a473af76-kube-api-access-9r2fd\") pod \"speaker-qb82t\" (UID: \"d2436da1-6ba8-4028-9012-1160a473af76\") " pod="metallb-system/speaker-qb82t" Feb 18 15:07:26 crc kubenswrapper[4896]: E0218 15:07:26.399969 4896 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 18 15:07:26 crc kubenswrapper[4896]: E0218 15:07:26.400031 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d2436da1-6ba8-4028-9012-1160a473af76-metrics-certs podName:d2436da1-6ba8-4028-9012-1160a473af76 nodeName:}" failed. No retries permitted until 2026-02-18 15:07:26.90000789 +0000 UTC m=+745.829342103 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d2436da1-6ba8-4028-9012-1160a473af76-metrics-certs") pod "speaker-qb82t" (UID: "d2436da1-6ba8-4028-9012-1160a473af76") : secret "speaker-certs-secret" not found Feb 18 15:07:26 crc kubenswrapper[4896]: E0218 15:07:26.400048 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d2436da1-6ba8-4028-9012-1160a473af76-memberlist podName:d2436da1-6ba8-4028-9012-1160a473af76 nodeName:}" failed. No retries permitted until 2026-02-18 15:07:26.900041341 +0000 UTC m=+745.829375554 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/d2436da1-6ba8-4028-9012-1160a473af76-memberlist") pod "speaker-qb82t" (UID: "d2436da1-6ba8-4028-9012-1160a473af76") : secret "metallb-memberlist" not found Feb 18 15:07:26 crc kubenswrapper[4896]: E0218 15:07:26.400067 4896 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Feb 18 15:07:26 crc kubenswrapper[4896]: E0218 15:07:26.400105 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7d817dfd-f711-48b3-8630-83d4b1a784de-metrics-certs podName:7d817dfd-f711-48b3-8630-83d4b1a784de nodeName:}" failed. No retries permitted until 2026-02-18 15:07:26.900082872 +0000 UTC m=+745.829417075 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7d817dfd-f711-48b3-8630-83d4b1a784de-metrics-certs") pod "controller-69bbfbf88f-zqgw4" (UID: "7d817dfd-f711-48b3-8630-83d4b1a784de") : secret "controller-certs-secret" not found Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.400574 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d2436da1-6ba8-4028-9012-1160a473af76-metallb-excludel2\") pod \"speaker-qb82t\" (UID: \"d2436da1-6ba8-4028-9012-1160a473af76\") " pod="metallb-system/speaker-qb82t" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.403131 4896 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.413223 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7d817dfd-f711-48b3-8630-83d4b1a784de-cert\") pod \"controller-69bbfbf88f-zqgw4\" (UID: \"7d817dfd-f711-48b3-8630-83d4b1a784de\") " pod="metallb-system/controller-69bbfbf88f-zqgw4" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.418693 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlgx6\" (UniqueName: \"kubernetes.io/projected/7d817dfd-f711-48b3-8630-83d4b1a784de-kube-api-access-qlgx6\") pod \"controller-69bbfbf88f-zqgw4\" (UID: \"7d817dfd-f711-48b3-8630-83d4b1a784de\") " pod="metallb-system/controller-69bbfbf88f-zqgw4" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.439751 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9r2fd\" (UniqueName: \"kubernetes.io/projected/d2436da1-6ba8-4028-9012-1160a473af76-kube-api-access-9r2fd\") pod \"speaker-qb82t\" (UID: \"d2436da1-6ba8-4028-9012-1160a473af76\") " pod="metallb-system/speaker-qb82t" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.477853 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-jxgkk" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.492314 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-lnrk2" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.698621 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-78b44bf5bb-jxgkk"] Feb 18 15:07:26 crc kubenswrapper[4896]: W0218 15:07:26.702314 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2a8e7f2_bf46_4f7d_a7e0_c1373f5c2136.slice/crio-db412ece686502dd48e8419456dd6c0ed4680ece3dcb101d097bae7f51a9ca94 WatchSource:0}: Error finding container db412ece686502dd48e8419456dd6c0ed4680ece3dcb101d097bae7f51a9ca94: Status 404 returned error can't find the container with id db412ece686502dd48e8419456dd6c0ed4680ece3dcb101d097bae7f51a9ca94 Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.911400 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d2436da1-6ba8-4028-9012-1160a473af76-metrics-certs\") pod \"speaker-qb82t\" (UID: \"d2436da1-6ba8-4028-9012-1160a473af76\") " pod="metallb-system/speaker-qb82t" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.911441 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d2436da1-6ba8-4028-9012-1160a473af76-memberlist\") pod \"speaker-qb82t\" (UID: \"d2436da1-6ba8-4028-9012-1160a473af76\") " pod="metallb-system/speaker-qb82t" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.911488 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7d817dfd-f711-48b3-8630-83d4b1a784de-metrics-certs\") pod \"controller-69bbfbf88f-zqgw4\" (UID: \"7d817dfd-f711-48b3-8630-83d4b1a784de\") " pod="metallb-system/controller-69bbfbf88f-zqgw4" Feb 18 15:07:26 crc kubenswrapper[4896]: E0218 15:07:26.912114 4896 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 18 15:07:26 crc kubenswrapper[4896]: E0218 15:07:26.912196 4896 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d2436da1-6ba8-4028-9012-1160a473af76-memberlist podName:d2436da1-6ba8-4028-9012-1160a473af76 nodeName:}" failed. No retries permitted until 2026-02-18 15:07:27.912177338 +0000 UTC m=+746.841511551 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/d2436da1-6ba8-4028-9012-1160a473af76-memberlist") pod "speaker-qb82t" (UID: "d2436da1-6ba8-4028-9012-1160a473af76") : secret "metallb-memberlist" not found Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.916132 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d2436da1-6ba8-4028-9012-1160a473af76-metrics-certs\") pod \"speaker-qb82t\" (UID: \"d2436da1-6ba8-4028-9012-1160a473af76\") " pod="metallb-system/speaker-qb82t" Feb 18 15:07:26 crc kubenswrapper[4896]: I0218 15:07:26.916312 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7d817dfd-f711-48b3-8630-83d4b1a784de-metrics-certs\") pod \"controller-69bbfbf88f-zqgw4\" (UID: \"7d817dfd-f711-48b3-8630-83d4b1a784de\") " pod="metallb-system/controller-69bbfbf88f-zqgw4" Feb 18 15:07:27 crc kubenswrapper[4896]: I0218 15:07:27.087105 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lnrk2" event={"ID":"29e01ca8-ce78-4769-a5d5-453834a27bb3","Type":"ContainerStarted","Data":"ce1e19c7d015212675965f369b67622b3845f7d9ed2a775a3b77867b9819d6ef"} Feb 18 15:07:27 crc kubenswrapper[4896]: I0218 15:07:27.088180 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-jxgkk" event={"ID":"c2a8e7f2-bf46-4f7d-a7e0-c1373f5c2136","Type":"ContainerStarted","Data":"db412ece686502dd48e8419456dd6c0ed4680ece3dcb101d097bae7f51a9ca94"} Feb 18 15:07:27 crc kubenswrapper[4896]: I0218 15:07:27.147188 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-69bbfbf88f-zqgw4" Feb 18 15:07:27 crc kubenswrapper[4896]: I0218 15:07:27.311846 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-69bbfbf88f-zqgw4"] Feb 18 15:07:27 crc kubenswrapper[4896]: W0218 15:07:27.320630 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d817dfd_f711_48b3_8630_83d4b1a784de.slice/crio-d43f9cbc624c4463bc37f98bacb3ad3e0bf9af30a316bf2810f407e2411f3f1c WatchSource:0}: Error finding container d43f9cbc624c4463bc37f98bacb3ad3e0bf9af30a316bf2810f407e2411f3f1c: Status 404 returned error can't find the container with id d43f9cbc624c4463bc37f98bacb3ad3e0bf9af30a316bf2810f407e2411f3f1c Feb 18 15:07:27 crc kubenswrapper[4896]: I0218 15:07:27.924655 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d2436da1-6ba8-4028-9012-1160a473af76-memberlist\") pod \"speaker-qb82t\" (UID: \"d2436da1-6ba8-4028-9012-1160a473af76\") " pod="metallb-system/speaker-qb82t" Feb 18 15:07:27 crc kubenswrapper[4896]: I0218 15:07:27.930171 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d2436da1-6ba8-4028-9012-1160a473af76-memberlist\") pod \"speaker-qb82t\" (UID: \"d2436da1-6ba8-4028-9012-1160a473af76\") " pod="metallb-system/speaker-qb82t" Feb 18 15:07:28 crc kubenswrapper[4896]: I0218 15:07:28.036964 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-qb82t" Feb 18 15:07:28 crc kubenswrapper[4896]: W0218 15:07:28.066593 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd2436da1_6ba8_4028_9012_1160a473af76.slice/crio-6bdfeade5718e208d2417f532264d40c32ebf1a1ce4d5014056e7880dee5528c WatchSource:0}: Error finding container 6bdfeade5718e208d2417f532264d40c32ebf1a1ce4d5014056e7880dee5528c: Status 404 returned error can't find the container with id 6bdfeade5718e208d2417f532264d40c32ebf1a1ce4d5014056e7880dee5528c Feb 18 15:07:28 crc kubenswrapper[4896]: I0218 15:07:28.098833 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-zqgw4" event={"ID":"7d817dfd-f711-48b3-8630-83d4b1a784de","Type":"ContainerStarted","Data":"8a1fc23f575c8ef36fd554184b4ad980010703160a84f8ffbdc38a105bc6a447"} Feb 18 15:07:28 crc kubenswrapper[4896]: I0218 15:07:28.099125 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-zqgw4" event={"ID":"7d817dfd-f711-48b3-8630-83d4b1a784de","Type":"ContainerStarted","Data":"d43f9cbc624c4463bc37f98bacb3ad3e0bf9af30a316bf2810f407e2411f3f1c"} Feb 18 15:07:28 crc kubenswrapper[4896]: I0218 15:07:28.100346 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qb82t" event={"ID":"d2436da1-6ba8-4028-9012-1160a473af76","Type":"ContainerStarted","Data":"6bdfeade5718e208d2417f532264d40c32ebf1a1ce4d5014056e7880dee5528c"} Feb 18 15:07:29 crc kubenswrapper[4896]: I0218 15:07:29.111087 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qb82t" event={"ID":"d2436da1-6ba8-4028-9012-1160a473af76","Type":"ContainerStarted","Data":"83576e175f3bf5bdbd469f21c9d025e8479e02adc54ee96181bc4356aaf0022e"} Feb 18 15:07:32 crc kubenswrapper[4896]: I0218 15:07:32.184187 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-69bbfbf88f-zqgw4" event={"ID":"7d817dfd-f711-48b3-8630-83d4b1a784de","Type":"ContainerStarted","Data":"94f12af540c523a6f09eebe0ac876b2c0b62106a1659149964f4426cb2f3ea3d"} Feb 18 15:07:32 crc kubenswrapper[4896]: I0218 15:07:32.184789 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-69bbfbf88f-zqgw4" Feb 18 15:07:32 crc kubenswrapper[4896]: I0218 15:07:32.186625 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-qb82t" event={"ID":"d2436da1-6ba8-4028-9012-1160a473af76","Type":"ContainerStarted","Data":"6e6425466959cb5babd5f6d889c3304695aa0092fe67b93c8d0e4b35b5cf9d15"} Feb 18 15:07:32 crc kubenswrapper[4896]: I0218 15:07:32.187253 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-qb82t" Feb 18 15:07:32 crc kubenswrapper[4896]: I0218 15:07:32.202465 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-69bbfbf88f-zqgw4" podStartSLOduration=2.586687419 podStartE2EDuration="6.202432843s" podCreationTimestamp="2026-02-18 15:07:26 +0000 UTC" firstStartedPulling="2026-02-18 15:07:27.425892898 +0000 UTC m=+746.355227111" lastFinishedPulling="2026-02-18 15:07:31.041638322 +0000 UTC m=+749.970972535" observedRunningTime="2026-02-18 15:07:32.199551065 +0000 UTC m=+751.128885278" watchObservedRunningTime="2026-02-18 15:07:32.202432843 +0000 UTC m=+751.131767056" Feb 18 15:07:32 crc kubenswrapper[4896]: I0218 15:07:32.216170 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-qb82t" podStartSLOduration=3.489834044 podStartE2EDuration="6.216153422s" podCreationTimestamp="2026-02-18 15:07:26 +0000 UTC" firstStartedPulling="2026-02-18 15:07:28.332774623 +0000 UTC m=+747.262108836" lastFinishedPulling="2026-02-18 15:07:31.059094001 +0000 UTC m=+749.988428214" observedRunningTime="2026-02-18 15:07:32.211589009 +0000 UTC m=+751.140923222" watchObservedRunningTime="2026-02-18 15:07:32.216153422 +0000 UTC m=+751.145487625" Feb 18 15:07:34 crc kubenswrapper[4896]: I0218 15:07:34.075672 4896 patch_prober.go:28] interesting pod/machine-config-daemon-24rq5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:07:34 crc kubenswrapper[4896]: I0218 15:07:34.076012 4896 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:07:34 crc kubenswrapper[4896]: I0218 15:07:34.199013 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-jxgkk" event={"ID":"c2a8e7f2-bf46-4f7d-a7e0-c1373f5c2136","Type":"ContainerStarted","Data":"50249051ea03938e23ac81ca7789fe23332eec794810afaefb8d530f788207d4"} Feb 18 15:07:34 crc kubenswrapper[4896]: I0218 15:07:34.199287 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-jxgkk" Feb 18 15:07:34 crc kubenswrapper[4896]: I0218 15:07:34.202039 4896 generic.go:334] "Generic (PLEG): container finished" podID="29e01ca8-ce78-4769-a5d5-453834a27bb3" containerID="b4fdb118d13342f070871b5243737d9295240773f9b2c2dd1d7f1c2ed1304f5d" exitCode=0 Feb 18 15:07:34 crc kubenswrapper[4896]: I0218 15:07:34.202152 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lnrk2" event={"ID":"29e01ca8-ce78-4769-a5d5-453834a27bb3","Type":"ContainerDied","Data":"b4fdb118d13342f070871b5243737d9295240773f9b2c2dd1d7f1c2ed1304f5d"} Feb 18 15:07:34 crc kubenswrapper[4896]: I0218 15:07:34.217392 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-jxgkk" podStartSLOduration=1.185212258 podStartE2EDuration="8.217376859s" podCreationTimestamp="2026-02-18 15:07:26 +0000 UTC" firstStartedPulling="2026-02-18 15:07:26.704499247 +0000 UTC m=+745.633833460" lastFinishedPulling="2026-02-18 15:07:33.736663848 +0000 UTC m=+752.665998061" observedRunningTime="2026-02-18 15:07:34.213926286 +0000 UTC m=+753.143260509" watchObservedRunningTime="2026-02-18 15:07:34.217376859 +0000 UTC m=+753.146711072" Feb 18 15:07:35 crc kubenswrapper[4896]: I0218 15:07:35.209379 4896 generic.go:334] "Generic (PLEG): container finished" podID="29e01ca8-ce78-4769-a5d5-453834a27bb3" containerID="2898d4457b19587b7962070a5bd0eede897d37c2791d1f7272e5afb50fe0c6a1" exitCode=0 Feb 18 15:07:35 crc kubenswrapper[4896]: I0218 15:07:35.209477 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lnrk2" event={"ID":"29e01ca8-ce78-4769-a5d5-453834a27bb3","Type":"ContainerDied","Data":"2898d4457b19587b7962070a5bd0eede897d37c2791d1f7272e5afb50fe0c6a1"} Feb 18 15:07:36 crc kubenswrapper[4896]: I0218 15:07:36.135817 4896 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 18 15:07:36 crc kubenswrapper[4896]: I0218 15:07:36.217670 4896 generic.go:334] "Generic (PLEG): container finished" podID="29e01ca8-ce78-4769-a5d5-453834a27bb3" containerID="2e137c6e3206eeadc563990d1d3ea0f3726529794bbfce0f47bd9c1f7dc9cee5" exitCode=0 Feb 18 15:07:36 crc kubenswrapper[4896]: I0218 15:07:36.217708 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lnrk2" event={"ID":"29e01ca8-ce78-4769-a5d5-453834a27bb3","Type":"ContainerDied","Data":"2e137c6e3206eeadc563990d1d3ea0f3726529794bbfce0f47bd9c1f7dc9cee5"} Feb 18 15:07:37 crc kubenswrapper[4896]: I0218 15:07:37.150571 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-69bbfbf88f-zqgw4" Feb 18 15:07:37 crc kubenswrapper[4896]: I0218 15:07:37.226632 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lnrk2" event={"ID":"29e01ca8-ce78-4769-a5d5-453834a27bb3","Type":"ContainerStarted","Data":"28869078a9ff710c8290ea262228c31fbd09a8cc17300bbd3649fcac0d10d7b3"} Feb 18 15:07:37 crc kubenswrapper[4896]: I0218 15:07:37.227328 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-lnrk2" Feb 18 15:07:37 crc kubenswrapper[4896]: I0218 15:07:37.227375 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lnrk2" event={"ID":"29e01ca8-ce78-4769-a5d5-453834a27bb3","Type":"ContainerStarted","Data":"af6e9ecfa49961f2d4806f45aa883d38620fb98333dbfd0a73d32c363bbea1d2"} Feb 18 15:07:37 crc kubenswrapper[4896]: I0218 15:07:37.227392 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lnrk2" event={"ID":"29e01ca8-ce78-4769-a5d5-453834a27bb3","Type":"ContainerStarted","Data":"1ce7f89c2ac69d439cb74ab037006767796dde33fdf7eb1d1b74915ce094ab2f"} Feb 18 15:07:37 crc kubenswrapper[4896]: I0218 15:07:37.227401 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lnrk2" event={"ID":"29e01ca8-ce78-4769-a5d5-453834a27bb3","Type":"ContainerStarted","Data":"d9a987668852e70d86b92404a515119018418fb1dc5674cb6e9727ec28553e5f"} Feb 18 15:07:37 crc kubenswrapper[4896]: I0218 15:07:37.227409 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lnrk2" event={"ID":"29e01ca8-ce78-4769-a5d5-453834a27bb3","Type":"ContainerStarted","Data":"4424318f379df2da3b659fb350150cccb46b05e003b60f76bc5dcc8449769984"} Feb 18 15:07:37 crc kubenswrapper[4896]: I0218 15:07:37.227417 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lnrk2" event={"ID":"29e01ca8-ce78-4769-a5d5-453834a27bb3","Type":"ContainerStarted","Data":"5b68b823a694a2768916e319856e539d493b5d238b81ed8688eeca382256a20e"} Feb 18 15:07:37 crc kubenswrapper[4896]: I0218 15:07:37.248859 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-lnrk2" podStartSLOduration=4.099461747 podStartE2EDuration="11.248839603s" podCreationTimestamp="2026-02-18 15:07:26 +0000 UTC" firstStartedPulling="2026-02-18 15:07:26.598692109 +0000 UTC m=+745.528026322" lastFinishedPulling="2026-02-18 15:07:33.748069955 +0000 UTC m=+752.677404178" observedRunningTime="2026-02-18 15:07:37.244055704 +0000 UTC m=+756.173389917" watchObservedRunningTime="2026-02-18 15:07:37.248839603 +0000 UTC m=+756.178173816" Feb 18 15:07:38 crc kubenswrapper[4896]: I0218 15:07:38.040682 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-qb82t" Feb 18 15:07:41 crc kubenswrapper[4896]: I0218 15:07:41.493292 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-lnrk2" Feb 18 15:07:41 crc kubenswrapper[4896]: I0218 15:07:41.547588 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-lnrk2" Feb 18 15:07:43 crc kubenswrapper[4896]: I0218 15:07:43.713412 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-lqkgg"] Feb 18 15:07:43 crc kubenswrapper[4896]: I0218 15:07:43.714564 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-lqkgg" Feb 18 15:07:43 crc kubenswrapper[4896]: I0218 15:07:43.716334 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Feb 18 15:07:43 crc kubenswrapper[4896]: I0218 15:07:43.716511 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-index-dockercfg-9tms4" Feb 18 15:07:43 crc kubenswrapper[4896]: I0218 15:07:43.716636 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Feb 18 15:07:43 crc kubenswrapper[4896]: I0218 15:07:43.728455 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-lqkgg"] Feb 18 15:07:43 crc kubenswrapper[4896]: I0218 15:07:43.840634 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcq7c\" (UniqueName: \"kubernetes.io/projected/32c638c3-ec07-42c1-9473-8f0483c5de4c-kube-api-access-xcq7c\") pod \"mariadb-operator-index-lqkgg\" (UID: \"32c638c3-ec07-42c1-9473-8f0483c5de4c\") " pod="openstack-operators/mariadb-operator-index-lqkgg" Feb 18 15:07:43 crc kubenswrapper[4896]: I0218 15:07:43.941306 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcq7c\" (UniqueName: \"kubernetes.io/projected/32c638c3-ec07-42c1-9473-8f0483c5de4c-kube-api-access-xcq7c\") pod \"mariadb-operator-index-lqkgg\" (UID: \"32c638c3-ec07-42c1-9473-8f0483c5de4c\") " pod="openstack-operators/mariadb-operator-index-lqkgg" Feb 18 15:07:43 crc kubenswrapper[4896]: I0218 15:07:43.958564 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcq7c\" (UniqueName: \"kubernetes.io/projected/32c638c3-ec07-42c1-9473-8f0483c5de4c-kube-api-access-xcq7c\") pod \"mariadb-operator-index-lqkgg\" (UID: \"32c638c3-ec07-42c1-9473-8f0483c5de4c\") " pod="openstack-operators/mariadb-operator-index-lqkgg" Feb 18 15:07:44 crc kubenswrapper[4896]: I0218 15:07:44.030761 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-lqkgg" Feb 18 15:07:44 crc kubenswrapper[4896]: I0218 15:07:44.453936 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-lqkgg"] Feb 18 15:07:44 crc kubenswrapper[4896]: W0218 15:07:44.457880 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod32c638c3_ec07_42c1_9473_8f0483c5de4c.slice/crio-08a7c81197cba6d1180b524fc1d4cee48950cedb2ceb16a4f46dcf2a7766d461 WatchSource:0}: Error finding container 08a7c81197cba6d1180b524fc1d4cee48950cedb2ceb16a4f46dcf2a7766d461: Status 404 returned error can't find the container with id 08a7c81197cba6d1180b524fc1d4cee48950cedb2ceb16a4f46dcf2a7766d461 Feb 18 15:07:45 crc kubenswrapper[4896]: I0218 15:07:45.265156 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-lqkgg" event={"ID":"32c638c3-ec07-42c1-9473-8f0483c5de4c","Type":"ContainerStarted","Data":"08a7c81197cba6d1180b524fc1d4cee48950cedb2ceb16a4f46dcf2a7766d461"} Feb 18 15:07:46 crc kubenswrapper[4896]: I0218 15:07:46.273394 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-lqkgg" event={"ID":"32c638c3-ec07-42c1-9473-8f0483c5de4c","Type":"ContainerStarted","Data":"15d665f228ccfc6a0ef3e11cdb1b6b857c2d187ea29d688b0c2dc9510db54560"} Feb 18 15:07:46 crc kubenswrapper[4896]: I0218 15:07:46.287860 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-lqkgg" podStartSLOduration=2.407790249 podStartE2EDuration="3.287844081s" podCreationTimestamp="2026-02-18 15:07:43 +0000 UTC" firstStartedPulling="2026-02-18 15:07:44.459695254 +0000 UTC m=+763.389029467" lastFinishedPulling="2026-02-18 15:07:45.339749086 +0000 UTC m=+764.269083299" observedRunningTime="2026-02-18 15:07:46.285320363 +0000 UTC m=+765.214654576" watchObservedRunningTime="2026-02-18 15:07:46.287844081 +0000 UTC m=+765.217178304" Feb 18 15:07:46 crc kubenswrapper[4896]: I0218 15:07:46.486894 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-78b44bf5bb-jxgkk" Feb 18 15:07:46 crc kubenswrapper[4896]: I0218 15:07:46.495912 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-lnrk2" Feb 18 15:07:54 crc kubenswrapper[4896]: I0218 15:07:54.031621 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-index-lqkgg" Feb 18 15:07:54 crc kubenswrapper[4896]: I0218 15:07:54.032864 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/mariadb-operator-index-lqkgg" Feb 18 15:07:54 crc kubenswrapper[4896]: I0218 15:07:54.060139 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/mariadb-operator-index-lqkgg" Feb 18 15:07:54 crc kubenswrapper[4896]: I0218 15:07:54.342382 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-index-lqkgg" Feb 18 15:08:01 crc kubenswrapper[4896]: I0218 15:08:01.697444 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt"] Feb 18 15:08:01 crc kubenswrapper[4896]: I0218 15:08:01.699677 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt" Feb 18 15:08:01 crc kubenswrapper[4896]: I0218 15:08:01.702700 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-9lxwl" Feb 18 15:08:01 crc kubenswrapper[4896]: I0218 15:08:01.706209 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt"] Feb 18 15:08:01 crc kubenswrapper[4896]: I0218 15:08:01.755728 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/87e452f6-c88e-4997-81e9-cab5976ab979-util\") pod \"4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt\" (UID: \"87e452f6-c88e-4997-81e9-cab5976ab979\") " pod="openstack-operators/4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt" Feb 18 15:08:01 crc kubenswrapper[4896]: I0218 15:08:01.756038 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4l5d2\" (UniqueName: \"kubernetes.io/projected/87e452f6-c88e-4997-81e9-cab5976ab979-kube-api-access-4l5d2\") pod \"4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt\" (UID: \"87e452f6-c88e-4997-81e9-cab5976ab979\") " pod="openstack-operators/4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt" Feb 18 15:08:01 crc kubenswrapper[4896]: I0218 15:08:01.756077 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/87e452f6-c88e-4997-81e9-cab5976ab979-bundle\") pod \"4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt\" (UID: \"87e452f6-c88e-4997-81e9-cab5976ab979\") " pod="openstack-operators/4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt" Feb 18 15:08:01 crc kubenswrapper[4896]: I0218 15:08:01.857143 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/87e452f6-c88e-4997-81e9-cab5976ab979-bundle\") pod \"4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt\" (UID: \"87e452f6-c88e-4997-81e9-cab5976ab979\") " pod="openstack-operators/4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt" Feb 18 15:08:01 crc kubenswrapper[4896]: I0218 15:08:01.857309 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/87e452f6-c88e-4997-81e9-cab5976ab979-util\") pod \"4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt\" (UID: \"87e452f6-c88e-4997-81e9-cab5976ab979\") " pod="openstack-operators/4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt" Feb 18 15:08:01 crc kubenswrapper[4896]: I0218 15:08:01.857333 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4l5d2\" (UniqueName: \"kubernetes.io/projected/87e452f6-c88e-4997-81e9-cab5976ab979-kube-api-access-4l5d2\") pod \"4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt\" (UID: \"87e452f6-c88e-4997-81e9-cab5976ab979\") " pod="openstack-operators/4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt" Feb 18 15:08:01 crc kubenswrapper[4896]: I0218 15:08:01.857704 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/87e452f6-c88e-4997-81e9-cab5976ab979-bundle\") pod \"4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt\" (UID: \"87e452f6-c88e-4997-81e9-cab5976ab979\") " pod="openstack-operators/4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt" Feb 18 15:08:01 crc kubenswrapper[4896]: I0218 15:08:01.857791 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/87e452f6-c88e-4997-81e9-cab5976ab979-util\") pod \"4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt\" (UID: \"87e452f6-c88e-4997-81e9-cab5976ab979\") " pod="openstack-operators/4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt" Feb 18 15:08:01 crc kubenswrapper[4896]: I0218 15:08:01.876178 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4l5d2\" (UniqueName: \"kubernetes.io/projected/87e452f6-c88e-4997-81e9-cab5976ab979-kube-api-access-4l5d2\") pod \"4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt\" (UID: \"87e452f6-c88e-4997-81e9-cab5976ab979\") " pod="openstack-operators/4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt" Feb 18 15:08:02 crc kubenswrapper[4896]: I0218 15:08:02.017719 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-9lxwl" Feb 18 15:08:02 crc kubenswrapper[4896]: I0218 15:08:02.026697 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt" Feb 18 15:08:02 crc kubenswrapper[4896]: I0218 15:08:02.193659 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt"] Feb 18 15:08:02 crc kubenswrapper[4896]: W0218 15:08:02.199542 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87e452f6_c88e_4997_81e9_cab5976ab979.slice/crio-c0f7d600ddff7ad8f0846ea8d36f60ee7ad4255366dd9fa728dee6c67eb3f0e2 WatchSource:0}: Error finding container c0f7d600ddff7ad8f0846ea8d36f60ee7ad4255366dd9fa728dee6c67eb3f0e2: Status 404 returned error can't find the container with id c0f7d600ddff7ad8f0846ea8d36f60ee7ad4255366dd9fa728dee6c67eb3f0e2 Feb 18 15:08:02 crc kubenswrapper[4896]: I0218 15:08:02.619981 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt" event={"ID":"87e452f6-c88e-4997-81e9-cab5976ab979","Type":"ContainerStarted","Data":"87f936fad317ae1428a0bccc7b8252bea2e7794b2a4f41323df3d09ff5cd021c"} Feb 18 15:08:02 crc kubenswrapper[4896]: I0218 15:08:02.620038 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt" event={"ID":"87e452f6-c88e-4997-81e9-cab5976ab979","Type":"ContainerStarted","Data":"c0f7d600ddff7ad8f0846ea8d36f60ee7ad4255366dd9fa728dee6c67eb3f0e2"} Feb 18 15:08:03 crc kubenswrapper[4896]: I0218 15:08:03.627801 4896 generic.go:334] "Generic (PLEG): container finished" podID="87e452f6-c88e-4997-81e9-cab5976ab979" containerID="87f936fad317ae1428a0bccc7b8252bea2e7794b2a4f41323df3d09ff5cd021c" exitCode=0 Feb 18 15:08:03 crc kubenswrapper[4896]: I0218 15:08:03.627914 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt" event={"ID":"87e452f6-c88e-4997-81e9-cab5976ab979","Type":"ContainerDied","Data":"87f936fad317ae1428a0bccc7b8252bea2e7794b2a4f41323df3d09ff5cd021c"} Feb 18 15:08:04 crc kubenswrapper[4896]: I0218 15:08:04.075998 4896 patch_prober.go:28] interesting pod/machine-config-daemon-24rq5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:08:04 crc kubenswrapper[4896]: I0218 15:08:04.076061 4896 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:08:04 crc kubenswrapper[4896]: I0218 15:08:04.076111 4896 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" Feb 18 15:08:04 crc kubenswrapper[4896]: I0218 15:08:04.076899 4896 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"91d6b5d30beb14a038ea309b143bbc7367c3406254b6275cdfed44c037d67a19"} pod="openshift-machine-config-operator/machine-config-daemon-24rq5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 18 15:08:04 crc kubenswrapper[4896]: I0218 15:08:04.076963 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" containerID="cri-o://91d6b5d30beb14a038ea309b143bbc7367c3406254b6275cdfed44c037d67a19" gracePeriod=600 Feb 18 15:08:04 crc kubenswrapper[4896]: I0218 15:08:04.634725 4896 generic.go:334] "Generic (PLEG): container finished" podID="87e452f6-c88e-4997-81e9-cab5976ab979" containerID="1c3e50bf8cd80b8282f334468b67199478ca5f7d76417f35bb27f562f4fb6098" exitCode=0 Feb 18 15:08:04 crc kubenswrapper[4896]: I0218 15:08:04.634906 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt" event={"ID":"87e452f6-c88e-4997-81e9-cab5976ab979","Type":"ContainerDied","Data":"1c3e50bf8cd80b8282f334468b67199478ca5f7d76417f35bb27f562f4fb6098"} Feb 18 15:08:04 crc kubenswrapper[4896]: I0218 15:08:04.637529 4896 generic.go:334] "Generic (PLEG): container finished" podID="8503734a-8111-4fab-beda-1f0f3b59615b" containerID="91d6b5d30beb14a038ea309b143bbc7367c3406254b6275cdfed44c037d67a19" exitCode=0 Feb 18 15:08:04 crc kubenswrapper[4896]: I0218 15:08:04.637557 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" event={"ID":"8503734a-8111-4fab-beda-1f0f3b59615b","Type":"ContainerDied","Data":"91d6b5d30beb14a038ea309b143bbc7367c3406254b6275cdfed44c037d67a19"} Feb 18 15:08:04 crc kubenswrapper[4896]: I0218 15:08:04.637590 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" event={"ID":"8503734a-8111-4fab-beda-1f0f3b59615b","Type":"ContainerStarted","Data":"763c989180a0be50e9c5a0c490b89d9b8795e413a592c48ba235f513fd0066bc"} Feb 18 15:08:04 crc kubenswrapper[4896]: I0218 15:08:04.637612 4896 scope.go:117] "RemoveContainer" containerID="e75d0500266e6b7af87e9b2cc3d752ab85959f8c74e00e1817c50ea02f263e92" Feb 18 15:08:05 crc kubenswrapper[4896]: I0218 15:08:05.644472 4896 generic.go:334] "Generic (PLEG): container finished" podID="87e452f6-c88e-4997-81e9-cab5976ab979" containerID="839f7e50eb64fcb6d73c1b52393660f7420d9cad3095d6e74daf7540befb25f4" exitCode=0 Feb 18 15:08:05 crc kubenswrapper[4896]: I0218 15:08:05.644540 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt" event={"ID":"87e452f6-c88e-4997-81e9-cab5976ab979","Type":"ContainerDied","Data":"839f7e50eb64fcb6d73c1b52393660f7420d9cad3095d6e74daf7540befb25f4"} Feb 18 15:08:06 crc kubenswrapper[4896]: I0218 15:08:06.896576 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt" Feb 18 15:08:06 crc kubenswrapper[4896]: I0218 15:08:06.917026 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/87e452f6-c88e-4997-81e9-cab5976ab979-bundle\") pod \"87e452f6-c88e-4997-81e9-cab5976ab979\" (UID: \"87e452f6-c88e-4997-81e9-cab5976ab979\") " Feb 18 15:08:06 crc kubenswrapper[4896]: I0218 15:08:06.917075 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4l5d2\" (UniqueName: \"kubernetes.io/projected/87e452f6-c88e-4997-81e9-cab5976ab979-kube-api-access-4l5d2\") pod \"87e452f6-c88e-4997-81e9-cab5976ab979\" (UID: \"87e452f6-c88e-4997-81e9-cab5976ab979\") " Feb 18 15:08:06 crc kubenswrapper[4896]: I0218 15:08:06.917100 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/87e452f6-c88e-4997-81e9-cab5976ab979-util\") pod \"87e452f6-c88e-4997-81e9-cab5976ab979\" (UID: \"87e452f6-c88e-4997-81e9-cab5976ab979\") " Feb 18 15:08:06 crc kubenswrapper[4896]: I0218 15:08:06.918265 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87e452f6-c88e-4997-81e9-cab5976ab979-bundle" (OuterVolumeSpecName: "bundle") pod "87e452f6-c88e-4997-81e9-cab5976ab979" (UID: "87e452f6-c88e-4997-81e9-cab5976ab979"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:08:06 crc kubenswrapper[4896]: I0218 15:08:06.922864 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87e452f6-c88e-4997-81e9-cab5976ab979-kube-api-access-4l5d2" (OuterVolumeSpecName: "kube-api-access-4l5d2") pod "87e452f6-c88e-4997-81e9-cab5976ab979" (UID: "87e452f6-c88e-4997-81e9-cab5976ab979"). InnerVolumeSpecName "kube-api-access-4l5d2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:08:06 crc kubenswrapper[4896]: I0218 15:08:06.929863 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87e452f6-c88e-4997-81e9-cab5976ab979-util" (OuterVolumeSpecName: "util") pod "87e452f6-c88e-4997-81e9-cab5976ab979" (UID: "87e452f6-c88e-4997-81e9-cab5976ab979"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:08:07 crc kubenswrapper[4896]: I0218 15:08:07.018608 4896 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/87e452f6-c88e-4997-81e9-cab5976ab979-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 15:08:07 crc kubenswrapper[4896]: I0218 15:08:07.018644 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4l5d2\" (UniqueName: \"kubernetes.io/projected/87e452f6-c88e-4997-81e9-cab5976ab979-kube-api-access-4l5d2\") on node \"crc\" DevicePath \"\"" Feb 18 15:08:07 crc kubenswrapper[4896]: I0218 15:08:07.018657 4896 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/87e452f6-c88e-4997-81e9-cab5976ab979-util\") on node \"crc\" DevicePath \"\"" Feb 18 15:08:07 crc kubenswrapper[4896]: I0218 15:08:07.658936 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt" event={"ID":"87e452f6-c88e-4997-81e9-cab5976ab979","Type":"ContainerDied","Data":"c0f7d600ddff7ad8f0846ea8d36f60ee7ad4255366dd9fa728dee6c67eb3f0e2"} Feb 18 15:08:07 crc kubenswrapper[4896]: I0218 15:08:07.658975 4896 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0f7d600ddff7ad8f0846ea8d36f60ee7ad4255366dd9fa728dee6c67eb3f0e2" Feb 18 15:08:07 crc kubenswrapper[4896]: I0218 15:08:07.658991 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt" Feb 18 15:08:15 crc kubenswrapper[4896]: I0218 15:08:15.199906 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-59746d8d7d-s27px"] Feb 18 15:08:15 crc kubenswrapper[4896]: E0218 15:08:15.200679 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87e452f6-c88e-4997-81e9-cab5976ab979" containerName="util" Feb 18 15:08:15 crc kubenswrapper[4896]: I0218 15:08:15.200694 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="87e452f6-c88e-4997-81e9-cab5976ab979" containerName="util" Feb 18 15:08:15 crc kubenswrapper[4896]: E0218 15:08:15.200707 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87e452f6-c88e-4997-81e9-cab5976ab979" containerName="extract" Feb 18 15:08:15 crc kubenswrapper[4896]: I0218 15:08:15.200715 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="87e452f6-c88e-4997-81e9-cab5976ab979" containerName="extract" Feb 18 15:08:15 crc kubenswrapper[4896]: E0218 15:08:15.200731 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87e452f6-c88e-4997-81e9-cab5976ab979" containerName="pull" Feb 18 15:08:15 crc kubenswrapper[4896]: I0218 15:08:15.200738 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="87e452f6-c88e-4997-81e9-cab5976ab979" containerName="pull" Feb 18 15:08:15 crc kubenswrapper[4896]: I0218 15:08:15.200868 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="87e452f6-c88e-4997-81e9-cab5976ab979" containerName="extract" Feb 18 15:08:15 crc kubenswrapper[4896]: I0218 15:08:15.201318 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-59746d8d7d-s27px" Feb 18 15:08:15 crc kubenswrapper[4896]: I0218 15:08:15.203182 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-wdc2j" Feb 18 15:08:15 crc kubenswrapper[4896]: I0218 15:08:15.203348 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Feb 18 15:08:15 crc kubenswrapper[4896]: I0218 15:08:15.204127 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-service-cert" Feb 18 15:08:15 crc kubenswrapper[4896]: I0218 15:08:15.214669 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-59746d8d7d-s27px"] Feb 18 15:08:15 crc kubenswrapper[4896]: I0218 15:08:15.368629 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rx996\" (UniqueName: \"kubernetes.io/projected/f1185065-2802-4e3e-a397-3170992e7b02-kube-api-access-rx996\") pod \"mariadb-operator-controller-manager-59746d8d7d-s27px\" (UID: \"f1185065-2802-4e3e-a397-3170992e7b02\") " pod="openstack-operators/mariadb-operator-controller-manager-59746d8d7d-s27px" Feb 18 15:08:15 crc kubenswrapper[4896]: I0218 15:08:15.368712 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f1185065-2802-4e3e-a397-3170992e7b02-webhook-cert\") pod \"mariadb-operator-controller-manager-59746d8d7d-s27px\" (UID: \"f1185065-2802-4e3e-a397-3170992e7b02\") " pod="openstack-operators/mariadb-operator-controller-manager-59746d8d7d-s27px" Feb 18 15:08:15 crc kubenswrapper[4896]: I0218 15:08:15.368738 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f1185065-2802-4e3e-a397-3170992e7b02-apiservice-cert\") pod \"mariadb-operator-controller-manager-59746d8d7d-s27px\" (UID: \"f1185065-2802-4e3e-a397-3170992e7b02\") " pod="openstack-operators/mariadb-operator-controller-manager-59746d8d7d-s27px" Feb 18 15:08:15 crc kubenswrapper[4896]: I0218 15:08:15.469715 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f1185065-2802-4e3e-a397-3170992e7b02-webhook-cert\") pod \"mariadb-operator-controller-manager-59746d8d7d-s27px\" (UID: \"f1185065-2802-4e3e-a397-3170992e7b02\") " pod="openstack-operators/mariadb-operator-controller-manager-59746d8d7d-s27px" Feb 18 15:08:15 crc kubenswrapper[4896]: I0218 15:08:15.469765 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f1185065-2802-4e3e-a397-3170992e7b02-apiservice-cert\") pod \"mariadb-operator-controller-manager-59746d8d7d-s27px\" (UID: \"f1185065-2802-4e3e-a397-3170992e7b02\") " pod="openstack-operators/mariadb-operator-controller-manager-59746d8d7d-s27px" Feb 18 15:08:15 crc kubenswrapper[4896]: I0218 15:08:15.469826 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rx996\" (UniqueName: \"kubernetes.io/projected/f1185065-2802-4e3e-a397-3170992e7b02-kube-api-access-rx996\") pod \"mariadb-operator-controller-manager-59746d8d7d-s27px\" (UID: \"f1185065-2802-4e3e-a397-3170992e7b02\") " pod="openstack-operators/mariadb-operator-controller-manager-59746d8d7d-s27px" Feb 18 15:08:15 crc kubenswrapper[4896]: I0218 15:08:15.475308 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f1185065-2802-4e3e-a397-3170992e7b02-apiservice-cert\") pod \"mariadb-operator-controller-manager-59746d8d7d-s27px\" (UID: \"f1185065-2802-4e3e-a397-3170992e7b02\") " pod="openstack-operators/mariadb-operator-controller-manager-59746d8d7d-s27px" Feb 18 15:08:15 crc kubenswrapper[4896]: I0218 15:08:15.475321 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f1185065-2802-4e3e-a397-3170992e7b02-webhook-cert\") pod \"mariadb-operator-controller-manager-59746d8d7d-s27px\" (UID: \"f1185065-2802-4e3e-a397-3170992e7b02\") " pod="openstack-operators/mariadb-operator-controller-manager-59746d8d7d-s27px" Feb 18 15:08:15 crc kubenswrapper[4896]: I0218 15:08:15.488192 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rx996\" (UniqueName: \"kubernetes.io/projected/f1185065-2802-4e3e-a397-3170992e7b02-kube-api-access-rx996\") pod \"mariadb-operator-controller-manager-59746d8d7d-s27px\" (UID: \"f1185065-2802-4e3e-a397-3170992e7b02\") " pod="openstack-operators/mariadb-operator-controller-manager-59746d8d7d-s27px" Feb 18 15:08:15 crc kubenswrapper[4896]: I0218 15:08:15.517940 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-59746d8d7d-s27px" Feb 18 15:08:15 crc kubenswrapper[4896]: I0218 15:08:15.722027 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-59746d8d7d-s27px"] Feb 18 15:08:16 crc kubenswrapper[4896]: I0218 15:08:16.716732 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-59746d8d7d-s27px" event={"ID":"f1185065-2802-4e3e-a397-3170992e7b02","Type":"ContainerStarted","Data":"181307a1445038ebd6d14f1f95e6f73b059107b51345419efc0161a8552335f5"} Feb 18 15:08:19 crc kubenswrapper[4896]: I0218 15:08:19.734744 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-59746d8d7d-s27px" event={"ID":"f1185065-2802-4e3e-a397-3170992e7b02","Type":"ContainerStarted","Data":"bc40715bb9e89af19778f5500a49c2414699e9de0d021a59c953140342819a55"} Feb 18 15:08:19 crc kubenswrapper[4896]: I0218 15:08:19.735294 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-59746d8d7d-s27px" Feb 18 15:08:19 crc kubenswrapper[4896]: I0218 15:08:19.750489 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-59746d8d7d-s27px" podStartSLOduration=1.8602729199999999 podStartE2EDuration="4.75046099s" podCreationTimestamp="2026-02-18 15:08:15 +0000 UTC" firstStartedPulling="2026-02-18 15:08:15.749850194 +0000 UTC m=+794.679184407" lastFinishedPulling="2026-02-18 15:08:18.640038264 +0000 UTC m=+797.569372477" observedRunningTime="2026-02-18 15:08:19.748541788 +0000 UTC m=+798.677876011" watchObservedRunningTime="2026-02-18 15:08:19.75046099 +0000 UTC m=+798.679795223" Feb 18 15:08:25 crc kubenswrapper[4896]: I0218 15:08:25.522071 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-59746d8d7d-s27px" Feb 18 15:08:27 crc kubenswrapper[4896]: I0218 15:08:27.767016 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-46bnc"] Feb 18 15:08:27 crc kubenswrapper[4896]: I0218 15:08:27.768606 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-46bnc" Feb 18 15:08:27 crc kubenswrapper[4896]: I0218 15:08:27.770538 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-index-dockercfg-nsmnx" Feb 18 15:08:27 crc kubenswrapper[4896]: I0218 15:08:27.773266 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-46bnc"] Feb 18 15:08:27 crc kubenswrapper[4896]: I0218 15:08:27.831752 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glvzq\" (UniqueName: \"kubernetes.io/projected/51243d7d-a0a1-4e05-8260-020300e92286-kube-api-access-glvzq\") pod \"infra-operator-index-46bnc\" (UID: \"51243d7d-a0a1-4e05-8260-020300e92286\") " pod="openstack-operators/infra-operator-index-46bnc" Feb 18 15:08:27 crc kubenswrapper[4896]: I0218 15:08:27.932698 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glvzq\" (UniqueName: \"kubernetes.io/projected/51243d7d-a0a1-4e05-8260-020300e92286-kube-api-access-glvzq\") pod \"infra-operator-index-46bnc\" (UID: \"51243d7d-a0a1-4e05-8260-020300e92286\") " pod="openstack-operators/infra-operator-index-46bnc" Feb 18 15:08:27 crc kubenswrapper[4896]: I0218 15:08:27.951466 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glvzq\" (UniqueName: \"kubernetes.io/projected/51243d7d-a0a1-4e05-8260-020300e92286-kube-api-access-glvzq\") pod \"infra-operator-index-46bnc\" (UID: \"51243d7d-a0a1-4e05-8260-020300e92286\") " pod="openstack-operators/infra-operator-index-46bnc" Feb 18 15:08:28 crc kubenswrapper[4896]: I0218 15:08:28.089018 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-46bnc" Feb 18 15:08:28 crc kubenswrapper[4896]: I0218 15:08:28.310701 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-46bnc"] Feb 18 15:08:28 crc kubenswrapper[4896]: I0218 15:08:28.781895 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-46bnc" event={"ID":"51243d7d-a0a1-4e05-8260-020300e92286","Type":"ContainerStarted","Data":"3ccb7ebc055ec3457eaffeabea6c59f121271ca0ffaa9a1cb92da12ddd736b67"} Feb 18 15:08:29 crc kubenswrapper[4896]: I0218 15:08:29.789045 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-46bnc" event={"ID":"51243d7d-a0a1-4e05-8260-020300e92286","Type":"ContainerStarted","Data":"76d148941b144239899b69c2779fe9d2f71f0b25f910bf67c8a4dd78f9f6c192"} Feb 18 15:08:29 crc kubenswrapper[4896]: I0218 15:08:29.806741 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-46bnc" podStartSLOduration=2.030662116 podStartE2EDuration="2.806721956s" podCreationTimestamp="2026-02-18 15:08:27 +0000 UTC" firstStartedPulling="2026-02-18 15:08:28.328966154 +0000 UTC m=+807.258300357" lastFinishedPulling="2026-02-18 15:08:29.105025984 +0000 UTC m=+808.034360197" observedRunningTime="2026-02-18 15:08:29.802122232 +0000 UTC m=+808.731456445" watchObservedRunningTime="2026-02-18 15:08:29.806721956 +0000 UTC m=+808.736056169" Feb 18 15:08:38 crc kubenswrapper[4896]: I0218 15:08:38.089930 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-index-46bnc" Feb 18 15:08:38 crc kubenswrapper[4896]: I0218 15:08:38.090492 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/infra-operator-index-46bnc" Feb 18 15:08:38 crc kubenswrapper[4896]: I0218 15:08:38.116462 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/infra-operator-index-46bnc" Feb 18 15:08:38 crc kubenswrapper[4896]: I0218 15:08:38.857476 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-index-46bnc" Feb 18 15:08:45 crc kubenswrapper[4896]: I0218 15:08:45.411959 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f"] Feb 18 15:08:45 crc kubenswrapper[4896]: I0218 15:08:45.414604 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f" Feb 18 15:08:45 crc kubenswrapper[4896]: I0218 15:08:45.416807 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-9lxwl" Feb 18 15:08:45 crc kubenswrapper[4896]: I0218 15:08:45.419843 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f"] Feb 18 15:08:45 crc kubenswrapper[4896]: I0218 15:08:45.538320 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/75e1f6d2-45d3-4805-9382-32d8a89ece0a-util\") pod \"97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f\" (UID: \"75e1f6d2-45d3-4805-9382-32d8a89ece0a\") " pod="openstack-operators/97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f" Feb 18 15:08:45 crc kubenswrapper[4896]: I0218 15:08:45.538567 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/75e1f6d2-45d3-4805-9382-32d8a89ece0a-bundle\") pod \"97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f\" (UID: \"75e1f6d2-45d3-4805-9382-32d8a89ece0a\") " pod="openstack-operators/97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f" Feb 18 15:08:45 crc kubenswrapper[4896]: I0218 15:08:45.538728 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xjr2\" (UniqueName: \"kubernetes.io/projected/75e1f6d2-45d3-4805-9382-32d8a89ece0a-kube-api-access-6xjr2\") pod \"97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f\" (UID: \"75e1f6d2-45d3-4805-9382-32d8a89ece0a\") " pod="openstack-operators/97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f" Feb 18 15:08:45 crc kubenswrapper[4896]: I0218 15:08:45.639752 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xjr2\" (UniqueName: \"kubernetes.io/projected/75e1f6d2-45d3-4805-9382-32d8a89ece0a-kube-api-access-6xjr2\") pod \"97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f\" (UID: \"75e1f6d2-45d3-4805-9382-32d8a89ece0a\") " pod="openstack-operators/97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f" Feb 18 15:08:45 crc kubenswrapper[4896]: I0218 15:08:45.639842 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/75e1f6d2-45d3-4805-9382-32d8a89ece0a-util\") pod \"97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f\" (UID: \"75e1f6d2-45d3-4805-9382-32d8a89ece0a\") " pod="openstack-operators/97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f" Feb 18 15:08:45 crc kubenswrapper[4896]: I0218 15:08:45.639918 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/75e1f6d2-45d3-4805-9382-32d8a89ece0a-bundle\") pod \"97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f\" (UID: \"75e1f6d2-45d3-4805-9382-32d8a89ece0a\") " pod="openstack-operators/97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f" Feb 18 15:08:45 crc kubenswrapper[4896]: I0218 15:08:45.640342 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/75e1f6d2-45d3-4805-9382-32d8a89ece0a-util\") pod \"97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f\" (UID: \"75e1f6d2-45d3-4805-9382-32d8a89ece0a\") " pod="openstack-operators/97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f" Feb 18 15:08:45 crc kubenswrapper[4896]: I0218 15:08:45.640557 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/75e1f6d2-45d3-4805-9382-32d8a89ece0a-bundle\") pod \"97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f\" (UID: \"75e1f6d2-45d3-4805-9382-32d8a89ece0a\") " pod="openstack-operators/97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f" Feb 18 15:08:45 crc kubenswrapper[4896]: I0218 15:08:45.659741 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xjr2\" (UniqueName: \"kubernetes.io/projected/75e1f6d2-45d3-4805-9382-32d8a89ece0a-kube-api-access-6xjr2\") pod \"97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f\" (UID: \"75e1f6d2-45d3-4805-9382-32d8a89ece0a\") " pod="openstack-operators/97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f" Feb 18 15:08:45 crc kubenswrapper[4896]: I0218 15:08:45.731959 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f" Feb 18 15:08:45 crc kubenswrapper[4896]: I0218 15:08:45.941162 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f"] Feb 18 15:08:45 crc kubenswrapper[4896]: W0218 15:08:45.957761 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75e1f6d2_45d3_4805_9382_32d8a89ece0a.slice/crio-f515b3411f809b36dc8985cf79c22c86f7af0d85e51ea6b2f7f8abf4abc495b9 WatchSource:0}: Error finding container f515b3411f809b36dc8985cf79c22c86f7af0d85e51ea6b2f7f8abf4abc495b9: Status 404 returned error can't find the container with id f515b3411f809b36dc8985cf79c22c86f7af0d85e51ea6b2f7f8abf4abc495b9 Feb 18 15:08:46 crc kubenswrapper[4896]: I0218 15:08:46.888652 4896 generic.go:334] "Generic (PLEG): container finished" podID="75e1f6d2-45d3-4805-9382-32d8a89ece0a" containerID="b4ce26bc5a344986b65b40a0257f901e2ca3286a39e1aefb6b4759998bbd07e0" exitCode=0 Feb 18 15:08:46 crc kubenswrapper[4896]: I0218 15:08:46.888727 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f" event={"ID":"75e1f6d2-45d3-4805-9382-32d8a89ece0a","Type":"ContainerDied","Data":"b4ce26bc5a344986b65b40a0257f901e2ca3286a39e1aefb6b4759998bbd07e0"} Feb 18 15:08:46 crc kubenswrapper[4896]: I0218 15:08:46.889047 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f" event={"ID":"75e1f6d2-45d3-4805-9382-32d8a89ece0a","Type":"ContainerStarted","Data":"f515b3411f809b36dc8985cf79c22c86f7af0d85e51ea6b2f7f8abf4abc495b9"} Feb 18 15:08:47 crc kubenswrapper[4896]: I0218 15:08:47.900067 4896 generic.go:334] "Generic (PLEG): container finished" podID="75e1f6d2-45d3-4805-9382-32d8a89ece0a" containerID="01a7e8403b441d814fee5451cd41e5f0160babf1a123cc50cd26e298e09b2687" exitCode=0 Feb 18 15:08:47 crc kubenswrapper[4896]: I0218 15:08:47.900106 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f" event={"ID":"75e1f6d2-45d3-4805-9382-32d8a89ece0a","Type":"ContainerDied","Data":"01a7e8403b441d814fee5451cd41e5f0160babf1a123cc50cd26e298e09b2687"} Feb 18 15:08:48 crc kubenswrapper[4896]: I0218 15:08:48.909619 4896 generic.go:334] "Generic (PLEG): container finished" podID="75e1f6d2-45d3-4805-9382-32d8a89ece0a" containerID="91a1e83ffc140a4dee2c4e8cecff0f679101337955f5ac50d7f9a792e47716f3" exitCode=0 Feb 18 15:08:48 crc kubenswrapper[4896]: I0218 15:08:48.909750 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f" event={"ID":"75e1f6d2-45d3-4805-9382-32d8a89ece0a","Type":"ContainerDied","Data":"91a1e83ffc140a4dee2c4e8cecff0f679101337955f5ac50d7f9a792e47716f3"} Feb 18 15:08:50 crc kubenswrapper[4896]: I0218 15:08:50.162132 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f" Feb 18 15:08:50 crc kubenswrapper[4896]: I0218 15:08:50.303703 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xjr2\" (UniqueName: \"kubernetes.io/projected/75e1f6d2-45d3-4805-9382-32d8a89ece0a-kube-api-access-6xjr2\") pod \"75e1f6d2-45d3-4805-9382-32d8a89ece0a\" (UID: \"75e1f6d2-45d3-4805-9382-32d8a89ece0a\") " Feb 18 15:08:50 crc kubenswrapper[4896]: I0218 15:08:50.303783 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/75e1f6d2-45d3-4805-9382-32d8a89ece0a-util\") pod \"75e1f6d2-45d3-4805-9382-32d8a89ece0a\" (UID: \"75e1f6d2-45d3-4805-9382-32d8a89ece0a\") " Feb 18 15:08:50 crc kubenswrapper[4896]: I0218 15:08:50.303818 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/75e1f6d2-45d3-4805-9382-32d8a89ece0a-bundle\") pod \"75e1f6d2-45d3-4805-9382-32d8a89ece0a\" (UID: \"75e1f6d2-45d3-4805-9382-32d8a89ece0a\") " Feb 18 15:08:50 crc kubenswrapper[4896]: I0218 15:08:50.305782 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75e1f6d2-45d3-4805-9382-32d8a89ece0a-bundle" (OuterVolumeSpecName: "bundle") pod "75e1f6d2-45d3-4805-9382-32d8a89ece0a" (UID: "75e1f6d2-45d3-4805-9382-32d8a89ece0a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:08:50 crc kubenswrapper[4896]: I0218 15:08:50.312475 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75e1f6d2-45d3-4805-9382-32d8a89ece0a-kube-api-access-6xjr2" (OuterVolumeSpecName: "kube-api-access-6xjr2") pod "75e1f6d2-45d3-4805-9382-32d8a89ece0a" (UID: "75e1f6d2-45d3-4805-9382-32d8a89ece0a"). InnerVolumeSpecName "kube-api-access-6xjr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:08:50 crc kubenswrapper[4896]: I0218 15:08:50.318753 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75e1f6d2-45d3-4805-9382-32d8a89ece0a-util" (OuterVolumeSpecName: "util") pod "75e1f6d2-45d3-4805-9382-32d8a89ece0a" (UID: "75e1f6d2-45d3-4805-9382-32d8a89ece0a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:08:50 crc kubenswrapper[4896]: I0218 15:08:50.409692 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xjr2\" (UniqueName: \"kubernetes.io/projected/75e1f6d2-45d3-4805-9382-32d8a89ece0a-kube-api-access-6xjr2\") on node \"crc\" DevicePath \"\"" Feb 18 15:08:50 crc kubenswrapper[4896]: I0218 15:08:50.409747 4896 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/75e1f6d2-45d3-4805-9382-32d8a89ece0a-util\") on node \"crc\" DevicePath \"\"" Feb 18 15:08:50 crc kubenswrapper[4896]: I0218 15:08:50.409763 4896 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/75e1f6d2-45d3-4805-9382-32d8a89ece0a-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 15:08:50 crc kubenswrapper[4896]: I0218 15:08:50.929761 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f" event={"ID":"75e1f6d2-45d3-4805-9382-32d8a89ece0a","Type":"ContainerDied","Data":"f515b3411f809b36dc8985cf79c22c86f7af0d85e51ea6b2f7f8abf4abc495b9"} Feb 18 15:08:50 crc kubenswrapper[4896]: I0218 15:08:50.929806 4896 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f515b3411f809b36dc8985cf79c22c86f7af0d85e51ea6b2f7f8abf4abc495b9" Feb 18 15:08:50 crc kubenswrapper[4896]: I0218 15:08:50.929810 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.347662 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["test-operator-kuttl-tests/openstack-galera-0"] Feb 18 15:09:00 crc kubenswrapper[4896]: E0218 15:09:00.348434 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75e1f6d2-45d3-4805-9382-32d8a89ece0a" containerName="pull" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.348451 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="75e1f6d2-45d3-4805-9382-32d8a89ece0a" containerName="pull" Feb 18 15:09:00 crc kubenswrapper[4896]: E0218 15:09:00.348469 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75e1f6d2-45d3-4805-9382-32d8a89ece0a" containerName="util" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.348477 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="75e1f6d2-45d3-4805-9382-32d8a89ece0a" containerName="util" Feb 18 15:09:00 crc kubenswrapper[4896]: E0218 15:09:00.348495 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75e1f6d2-45d3-4805-9382-32d8a89ece0a" containerName="extract" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.348504 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="75e1f6d2-45d3-4805-9382-32d8a89ece0a" containerName="extract" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.348620 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="75e1f6d2-45d3-4805-9382-32d8a89ece0a" containerName="extract" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.349353 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="test-operator-kuttl-tests/openstack-galera-0" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.352452 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"test-operator-kuttl-tests"/"openshift-service-ca.crt" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.353003 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"test-operator-kuttl-tests"/"openstack-config-data" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.353026 4896 reflector.go:368] Caches populated for *v1.Secret from object-"test-operator-kuttl-tests"/"galera-openstack-dockercfg-2nzmd" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.353061 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"test-operator-kuttl-tests"/"openstack-scripts" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.353520 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"test-operator-kuttl-tests"/"kube-root-ca.crt" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.365312 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["test-operator-kuttl-tests/openstack-galera-2"] Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.366212 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="test-operator-kuttl-tests/openstack-galera-2" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.368418 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["test-operator-kuttl-tests/openstack-galera-1"] Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.369454 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="test-operator-kuttl-tests/openstack-galera-1" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.372012 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["test-operator-kuttl-tests/openstack-galera-0"] Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.391474 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["test-operator-kuttl-tests/openstack-galera-1"] Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.395802 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["test-operator-kuttl-tests/openstack-galera-2"] Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.537180 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-588vr\" (UniqueName: \"kubernetes.io/projected/1a1927a6-afa4-41c7-95ad-22d7bba44d38-kube-api-access-588vr\") pod \"openstack-galera-1\" (UID: \"1a1927a6-afa4-41c7-95ad-22d7bba44d38\") " pod="test-operator-kuttl-tests/openstack-galera-1" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.537291 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed\") " pod="test-operator-kuttl-tests/openstack-galera-0" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.537327 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0e13f33c-acb4-421a-be84-4757ff7c3b74-config-data-default\") pod \"openstack-galera-2\" (UID: \"0e13f33c-acb4-421a-be84-4757ff7c3b74\") " pod="test-operator-kuttl-tests/openstack-galera-2" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.537376 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed-config-data-generated\") pod \"openstack-galera-0\" (UID: \"1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed\") " pod="test-operator-kuttl-tests/openstack-galera-0" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.537405 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jc6lp\" (UniqueName: \"kubernetes.io/projected/1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed-kube-api-access-jc6lp\") pod \"openstack-galera-0\" (UID: \"1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed\") " pod="test-operator-kuttl-tests/openstack-galera-0" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.537486 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1a1927a6-afa4-41c7-95ad-22d7bba44d38-config-data-default\") pod \"openstack-galera-1\" (UID: \"1a1927a6-afa4-41c7-95ad-22d7bba44d38\") " pod="test-operator-kuttl-tests/openstack-galera-1" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.537541 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed-kolla-config\") pod \"openstack-galera-0\" (UID: \"1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed\") " pod="test-operator-kuttl-tests/openstack-galera-0" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.537598 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-2\" (UID: \"0e13f33c-acb4-421a-be84-4757ff7c3b74\") " pod="test-operator-kuttl-tests/openstack-galera-2" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.537640 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-1\" (UID: \"1a1927a6-afa4-41c7-95ad-22d7bba44d38\") " pod="test-operator-kuttl-tests/openstack-galera-1" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.537659 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a1927a6-afa4-41c7-95ad-22d7bba44d38-operator-scripts\") pod \"openstack-galera-1\" (UID: \"1a1927a6-afa4-41c7-95ad-22d7bba44d38\") " pod="test-operator-kuttl-tests/openstack-galera-1" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.537683 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0e13f33c-acb4-421a-be84-4757ff7c3b74-kolla-config\") pod \"openstack-galera-2\" (UID: \"0e13f33c-acb4-421a-be84-4757ff7c3b74\") " pod="test-operator-kuttl-tests/openstack-galera-2" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.537711 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1a1927a6-afa4-41c7-95ad-22d7bba44d38-kolla-config\") pod \"openstack-galera-1\" (UID: \"1a1927a6-afa4-41c7-95ad-22d7bba44d38\") " pod="test-operator-kuttl-tests/openstack-galera-1" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.537727 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e13f33c-acb4-421a-be84-4757ff7c3b74-operator-scripts\") pod \"openstack-galera-2\" (UID: \"0e13f33c-acb4-421a-be84-4757ff7c3b74\") " pod="test-operator-kuttl-tests/openstack-galera-2" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.537763 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed-config-data-default\") pod \"openstack-galera-0\" (UID: \"1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed\") " pod="test-operator-kuttl-tests/openstack-galera-0" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.537782 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed-operator-scripts\") pod \"openstack-galera-0\" (UID: \"1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed\") " pod="test-operator-kuttl-tests/openstack-galera-0" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.537798 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztnhw\" (UniqueName: \"kubernetes.io/projected/0e13f33c-acb4-421a-be84-4757ff7c3b74-kube-api-access-ztnhw\") pod \"openstack-galera-2\" (UID: \"0e13f33c-acb4-421a-be84-4757ff7c3b74\") " pod="test-operator-kuttl-tests/openstack-galera-2" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.537820 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0e13f33c-acb4-421a-be84-4757ff7c3b74-config-data-generated\") pod \"openstack-galera-2\" (UID: \"0e13f33c-acb4-421a-be84-4757ff7c3b74\") " pod="test-operator-kuttl-tests/openstack-galera-2" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.537852 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1a1927a6-afa4-41c7-95ad-22d7bba44d38-config-data-generated\") pod \"openstack-galera-1\" (UID: \"1a1927a6-afa4-41c7-95ad-22d7bba44d38\") " pod="test-operator-kuttl-tests/openstack-galera-1" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.639475 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0e13f33c-acb4-421a-be84-4757ff7c3b74-config-data-default\") pod \"openstack-galera-2\" (UID: \"0e13f33c-acb4-421a-be84-4757ff7c3b74\") " pod="test-operator-kuttl-tests/openstack-galera-2" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.639543 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed-config-data-generated\") pod \"openstack-galera-0\" (UID: \"1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed\") " pod="test-operator-kuttl-tests/openstack-galera-0" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.639565 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jc6lp\" (UniqueName: \"kubernetes.io/projected/1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed-kube-api-access-jc6lp\") pod \"openstack-galera-0\" (UID: \"1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed\") " pod="test-operator-kuttl-tests/openstack-galera-0" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.639587 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1a1927a6-afa4-41c7-95ad-22d7bba44d38-config-data-default\") pod \"openstack-galera-1\" (UID: \"1a1927a6-afa4-41c7-95ad-22d7bba44d38\") " pod="test-operator-kuttl-tests/openstack-galera-1" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.639605 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed-kolla-config\") pod \"openstack-galera-0\" (UID: \"1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed\") " pod="test-operator-kuttl-tests/openstack-galera-0" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.639633 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-2\" (UID: \"0e13f33c-acb4-421a-be84-4757ff7c3b74\") " pod="test-operator-kuttl-tests/openstack-galera-2" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.639684 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-1\" (UID: \"1a1927a6-afa4-41c7-95ad-22d7bba44d38\") " pod="test-operator-kuttl-tests/openstack-galera-1" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.639706 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a1927a6-afa4-41c7-95ad-22d7bba44d38-operator-scripts\") pod \"openstack-galera-1\" (UID: \"1a1927a6-afa4-41c7-95ad-22d7bba44d38\") " pod="test-operator-kuttl-tests/openstack-galera-1" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.639734 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0e13f33c-acb4-421a-be84-4757ff7c3b74-kolla-config\") pod \"openstack-galera-2\" (UID: \"0e13f33c-acb4-421a-be84-4757ff7c3b74\") " pod="test-operator-kuttl-tests/openstack-galera-2" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.639783 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1a1927a6-afa4-41c7-95ad-22d7bba44d38-kolla-config\") pod \"openstack-galera-1\" (UID: \"1a1927a6-afa4-41c7-95ad-22d7bba44d38\") " pod="test-operator-kuttl-tests/openstack-galera-1" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.639806 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e13f33c-acb4-421a-be84-4757ff7c3b74-operator-scripts\") pod \"openstack-galera-2\" (UID: \"0e13f33c-acb4-421a-be84-4757ff7c3b74\") " pod="test-operator-kuttl-tests/openstack-galera-2" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.639830 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed-config-data-default\") pod \"openstack-galera-0\" (UID: \"1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed\") " pod="test-operator-kuttl-tests/openstack-galera-0" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.639852 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed-operator-scripts\") pod \"openstack-galera-0\" (UID: \"1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed\") " pod="test-operator-kuttl-tests/openstack-galera-0" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.639876 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztnhw\" (UniqueName: \"kubernetes.io/projected/0e13f33c-acb4-421a-be84-4757ff7c3b74-kube-api-access-ztnhw\") pod \"openstack-galera-2\" (UID: \"0e13f33c-acb4-421a-be84-4757ff7c3b74\") " pod="test-operator-kuttl-tests/openstack-galera-2" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.639898 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0e13f33c-acb4-421a-be84-4757ff7c3b74-config-data-generated\") pod \"openstack-galera-2\" (UID: \"0e13f33c-acb4-421a-be84-4757ff7c3b74\") " pod="test-operator-kuttl-tests/openstack-galera-2" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.639921 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1a1927a6-afa4-41c7-95ad-22d7bba44d38-config-data-generated\") pod \"openstack-galera-1\" (UID: \"1a1927a6-afa4-41c7-95ad-22d7bba44d38\") " pod="test-operator-kuttl-tests/openstack-galera-1" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.639949 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-588vr\" (UniqueName: \"kubernetes.io/projected/1a1927a6-afa4-41c7-95ad-22d7bba44d38-kube-api-access-588vr\") pod \"openstack-galera-1\" (UID: \"1a1927a6-afa4-41c7-95ad-22d7bba44d38\") " pod="test-operator-kuttl-tests/openstack-galera-1" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.639973 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed\") " pod="test-operator-kuttl-tests/openstack-galera-0" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.640272 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed-config-data-generated\") pod \"openstack-galera-0\" (UID: \"1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed\") " pod="test-operator-kuttl-tests/openstack-galera-0" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.640320 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0e13f33c-acb4-421a-be84-4757ff7c3b74-config-data-default\") pod \"openstack-galera-2\" (UID: \"0e13f33c-acb4-421a-be84-4757ff7c3b74\") " pod="test-operator-kuttl-tests/openstack-galera-2" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.640393 4896 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed\") device mount path \"/mnt/openstack/pv01\"" pod="test-operator-kuttl-tests/openstack-galera-0" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.640515 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1a1927a6-afa4-41c7-95ad-22d7bba44d38-config-data-default\") pod \"openstack-galera-1\" (UID: \"1a1927a6-afa4-41c7-95ad-22d7bba44d38\") " pod="test-operator-kuttl-tests/openstack-galera-1" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.641355 4896 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-2\" (UID: \"0e13f33c-acb4-421a-be84-4757ff7c3b74\") device mount path \"/mnt/openstack/pv02\"" pod="test-operator-kuttl-tests/openstack-galera-2" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.641411 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0e13f33c-acb4-421a-be84-4757ff7c3b74-kolla-config\") pod \"openstack-galera-2\" (UID: \"0e13f33c-acb4-421a-be84-4757ff7c3b74\") " pod="test-operator-kuttl-tests/openstack-galera-2" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.641496 4896 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-1\" (UID: \"1a1927a6-afa4-41c7-95ad-22d7bba44d38\") device mount path \"/mnt/openstack/pv09\"" pod="test-operator-kuttl-tests/openstack-galera-1" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.641544 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed-kolla-config\") pod \"openstack-galera-0\" (UID: \"1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed\") " pod="test-operator-kuttl-tests/openstack-galera-0" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.642356 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a1927a6-afa4-41c7-95ad-22d7bba44d38-operator-scripts\") pod \"openstack-galera-1\" (UID: \"1a1927a6-afa4-41c7-95ad-22d7bba44d38\") " pod="test-operator-kuttl-tests/openstack-galera-1" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.642657 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed-config-data-default\") pod \"openstack-galera-0\" (UID: \"1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed\") " pod="test-operator-kuttl-tests/openstack-galera-0" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.643400 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed-operator-scripts\") pod \"openstack-galera-0\" (UID: \"1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed\") " pod="test-operator-kuttl-tests/openstack-galera-0" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.645506 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0e13f33c-acb4-421a-be84-4757ff7c3b74-config-data-generated\") pod \"openstack-galera-2\" (UID: \"0e13f33c-acb4-421a-be84-4757ff7c3b74\") " pod="test-operator-kuttl-tests/openstack-galera-2" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.645871 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1a1927a6-afa4-41c7-95ad-22d7bba44d38-config-data-generated\") pod \"openstack-galera-1\" (UID: \"1a1927a6-afa4-41c7-95ad-22d7bba44d38\") " pod="test-operator-kuttl-tests/openstack-galera-1" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.645986 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1a1927a6-afa4-41c7-95ad-22d7bba44d38-kolla-config\") pod \"openstack-galera-1\" (UID: \"1a1927a6-afa4-41c7-95ad-22d7bba44d38\") " pod="test-operator-kuttl-tests/openstack-galera-1" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.646538 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0e13f33c-acb4-421a-be84-4757ff7c3b74-operator-scripts\") pod \"openstack-galera-2\" (UID: \"0e13f33c-acb4-421a-be84-4757ff7c3b74\") " pod="test-operator-kuttl-tests/openstack-galera-2" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.659960 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jc6lp\" (UniqueName: \"kubernetes.io/projected/1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed-kube-api-access-jc6lp\") pod \"openstack-galera-0\" (UID: \"1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed\") " pod="test-operator-kuttl-tests/openstack-galera-0" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.660006 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztnhw\" (UniqueName: \"kubernetes.io/projected/0e13f33c-acb4-421a-be84-4757ff7c3b74-kube-api-access-ztnhw\") pod \"openstack-galera-2\" (UID: \"0e13f33c-acb4-421a-be84-4757ff7c3b74\") " pod="test-operator-kuttl-tests/openstack-galera-2" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.661418 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-1\" (UID: \"1a1927a6-afa4-41c7-95ad-22d7bba44d38\") " pod="test-operator-kuttl-tests/openstack-galera-1" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.663639 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed\") " pod="test-operator-kuttl-tests/openstack-galera-0" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.665320 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-588vr\" (UniqueName: \"kubernetes.io/projected/1a1927a6-afa4-41c7-95ad-22d7bba44d38-kube-api-access-588vr\") pod \"openstack-galera-1\" (UID: \"1a1927a6-afa4-41c7-95ad-22d7bba44d38\") " pod="test-operator-kuttl-tests/openstack-galera-1" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.678401 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-2\" (UID: \"0e13f33c-acb4-421a-be84-4757ff7c3b74\") " pod="test-operator-kuttl-tests/openstack-galera-2" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.699762 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="test-operator-kuttl-tests/openstack-galera-1" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.758122 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-5d86b96dc9-dhl2n"] Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.759008 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5d86b96dc9-dhl2n" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.762948 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-service-cert" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.763225 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-p7wxt" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.770734 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5d86b96dc9-dhl2n"] Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.847764 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bf5a2769-d00a-4493-825f-006f6607e79b-webhook-cert\") pod \"infra-operator-controller-manager-5d86b96dc9-dhl2n\" (UID: \"bf5a2769-d00a-4493-825f-006f6607e79b\") " pod="openstack-operators/infra-operator-controller-manager-5d86b96dc9-dhl2n" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.847809 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bf5a2769-d00a-4493-825f-006f6607e79b-apiservice-cert\") pod \"infra-operator-controller-manager-5d86b96dc9-dhl2n\" (UID: \"bf5a2769-d00a-4493-825f-006f6607e79b\") " pod="openstack-operators/infra-operator-controller-manager-5d86b96dc9-dhl2n" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.847859 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kw6tk\" (UniqueName: \"kubernetes.io/projected/bf5a2769-d00a-4493-825f-006f6607e79b-kube-api-access-kw6tk\") pod \"infra-operator-controller-manager-5d86b96dc9-dhl2n\" (UID: \"bf5a2769-d00a-4493-825f-006f6607e79b\") " pod="openstack-operators/infra-operator-controller-manager-5d86b96dc9-dhl2n" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.948929 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kw6tk\" (UniqueName: \"kubernetes.io/projected/bf5a2769-d00a-4493-825f-006f6607e79b-kube-api-access-kw6tk\") pod \"infra-operator-controller-manager-5d86b96dc9-dhl2n\" (UID: \"bf5a2769-d00a-4493-825f-006f6607e79b\") " pod="openstack-operators/infra-operator-controller-manager-5d86b96dc9-dhl2n" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.949030 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bf5a2769-d00a-4493-825f-006f6607e79b-webhook-cert\") pod \"infra-operator-controller-manager-5d86b96dc9-dhl2n\" (UID: \"bf5a2769-d00a-4493-825f-006f6607e79b\") " pod="openstack-operators/infra-operator-controller-manager-5d86b96dc9-dhl2n" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.949061 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bf5a2769-d00a-4493-825f-006f6607e79b-apiservice-cert\") pod \"infra-operator-controller-manager-5d86b96dc9-dhl2n\" (UID: \"bf5a2769-d00a-4493-825f-006f6607e79b\") " pod="openstack-operators/infra-operator-controller-manager-5d86b96dc9-dhl2n" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.953786 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bf5a2769-d00a-4493-825f-006f6607e79b-apiservice-cert\") pod \"infra-operator-controller-manager-5d86b96dc9-dhl2n\" (UID: \"bf5a2769-d00a-4493-825f-006f6607e79b\") " pod="openstack-operators/infra-operator-controller-manager-5d86b96dc9-dhl2n" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.955718 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bf5a2769-d00a-4493-825f-006f6607e79b-webhook-cert\") pod \"infra-operator-controller-manager-5d86b96dc9-dhl2n\" (UID: \"bf5a2769-d00a-4493-825f-006f6607e79b\") " pod="openstack-operators/infra-operator-controller-manager-5d86b96dc9-dhl2n" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.963710 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="test-operator-kuttl-tests/openstack-galera-0" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.982446 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="test-operator-kuttl-tests/openstack-galera-2" Feb 18 15:09:00 crc kubenswrapper[4896]: I0218 15:09:00.982500 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kw6tk\" (UniqueName: \"kubernetes.io/projected/bf5a2769-d00a-4493-825f-006f6607e79b-kube-api-access-kw6tk\") pod \"infra-operator-controller-manager-5d86b96dc9-dhl2n\" (UID: \"bf5a2769-d00a-4493-825f-006f6607e79b\") " pod="openstack-operators/infra-operator-controller-manager-5d86b96dc9-dhl2n" Feb 18 15:09:01 crc kubenswrapper[4896]: I0218 15:09:01.094323 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5d86b96dc9-dhl2n" Feb 18 15:09:01 crc kubenswrapper[4896]: I0218 15:09:01.169379 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["test-operator-kuttl-tests/openstack-galera-1"] Feb 18 15:09:01 crc kubenswrapper[4896]: I0218 15:09:01.282918 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["test-operator-kuttl-tests/openstack-galera-2"] Feb 18 15:09:01 crc kubenswrapper[4896]: W0218 15:09:01.284684 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e13f33c_acb4_421a_be84_4757ff7c3b74.slice/crio-56ddd4e4eafe8d625cb5e538a73f3ff48bd9fbc22b5221e8a67c4b7e6c37d79a WatchSource:0}: Error finding container 56ddd4e4eafe8d625cb5e538a73f3ff48bd9fbc22b5221e8a67c4b7e6c37d79a: Status 404 returned error can't find the container with id 56ddd4e4eafe8d625cb5e538a73f3ff48bd9fbc22b5221e8a67c4b7e6c37d79a Feb 18 15:09:01 crc kubenswrapper[4896]: I0218 15:09:01.339697 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5d86b96dc9-dhl2n"] Feb 18 15:09:01 crc kubenswrapper[4896]: W0218 15:09:01.343446 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf5a2769_d00a_4493_825f_006f6607e79b.slice/crio-f131f1f5e7bf4d4897b1829a0672ae5d62e59b1a31865b36b5af0d79839df1c8 WatchSource:0}: Error finding container f131f1f5e7bf4d4897b1829a0672ae5d62e59b1a31865b36b5af0d79839df1c8: Status 404 returned error can't find the container with id f131f1f5e7bf4d4897b1829a0672ae5d62e59b1a31865b36b5af0d79839df1c8 Feb 18 15:09:01 crc kubenswrapper[4896]: I0218 15:09:01.428653 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["test-operator-kuttl-tests/openstack-galera-0"] Feb 18 15:09:01 crc kubenswrapper[4896]: W0218 15:09:01.434899 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1fb2a5b0_19cb_4d59_8d9d_13ffb1c499ed.slice/crio-135561870e4979d882b11018d6805740bd6f910faa8d10c09155a3bc75f38079 WatchSource:0}: Error finding container 135561870e4979d882b11018d6805740bd6f910faa8d10c09155a3bc75f38079: Status 404 returned error can't find the container with id 135561870e4979d882b11018d6805740bd6f910faa8d10c09155a3bc75f38079 Feb 18 15:09:02 crc kubenswrapper[4896]: I0218 15:09:02.019707 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5d86b96dc9-dhl2n" event={"ID":"bf5a2769-d00a-4493-825f-006f6607e79b","Type":"ContainerStarted","Data":"f131f1f5e7bf4d4897b1829a0672ae5d62e59b1a31865b36b5af0d79839df1c8"} Feb 18 15:09:02 crc kubenswrapper[4896]: I0218 15:09:02.035458 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/openstack-galera-2" event={"ID":"0e13f33c-acb4-421a-be84-4757ff7c3b74","Type":"ContainerStarted","Data":"56ddd4e4eafe8d625cb5e538a73f3ff48bd9fbc22b5221e8a67c4b7e6c37d79a"} Feb 18 15:09:02 crc kubenswrapper[4896]: I0218 15:09:02.042501 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/openstack-galera-0" event={"ID":"1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed","Type":"ContainerStarted","Data":"135561870e4979d882b11018d6805740bd6f910faa8d10c09155a3bc75f38079"} Feb 18 15:09:02 crc kubenswrapper[4896]: I0218 15:09:02.055385 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/openstack-galera-1" event={"ID":"1a1927a6-afa4-41c7-95ad-22d7bba44d38","Type":"ContainerStarted","Data":"d3f1b73afcdbccdf92253ad9ef6feb61997c4e9c48c07f4a4d59e475a84edc7d"} Feb 18 15:09:13 crc kubenswrapper[4896]: I0218 15:09:13.543461 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/openstack-galera-1" event={"ID":"1a1927a6-afa4-41c7-95ad-22d7bba44d38","Type":"ContainerStarted","Data":"a284c2e2fe2036ab6c2b4010bee31b3a50cd6c4aeb0284a6147f2ad000d0e748"} Feb 18 15:09:13 crc kubenswrapper[4896]: I0218 15:09:13.545732 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/openstack-galera-2" event={"ID":"0e13f33c-acb4-421a-be84-4757ff7c3b74","Type":"ContainerStarted","Data":"71e9ad2c7f8e1409e222a5a29e82e3c4795fa2d58deab669795283992f6413fd"} Feb 18 15:09:13 crc kubenswrapper[4896]: I0218 15:09:13.547104 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5d86b96dc9-dhl2n" event={"ID":"bf5a2769-d00a-4493-825f-006f6607e79b","Type":"ContainerStarted","Data":"0cb0253c717e3c13566fe6f9b8b143f70464f54cfba1657affa48569537e1493"} Feb 18 15:09:13 crc kubenswrapper[4896]: I0218 15:09:13.547251 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-5d86b96dc9-dhl2n" Feb 18 15:09:13 crc kubenswrapper[4896]: I0218 15:09:13.548326 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/openstack-galera-0" event={"ID":"1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed","Type":"ContainerStarted","Data":"5cfc8cd25d326b00c6e48ecc4a150c223599feef13c2ab0e697363415ff36dd9"} Feb 18 15:09:13 crc kubenswrapper[4896]: I0218 15:09:13.598571 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-5d86b96dc9-dhl2n" podStartSLOduration=1.974844007 podStartE2EDuration="13.598556271s" podCreationTimestamp="2026-02-18 15:09:00 +0000 UTC" firstStartedPulling="2026-02-18 15:09:01.346267587 +0000 UTC m=+840.275601800" lastFinishedPulling="2026-02-18 15:09:12.969979851 +0000 UTC m=+851.899314064" observedRunningTime="2026-02-18 15:09:13.596611659 +0000 UTC m=+852.525945872" watchObservedRunningTime="2026-02-18 15:09:13.598556271 +0000 UTC m=+852.527890484" Feb 18 15:09:17 crc kubenswrapper[4896]: I0218 15:09:17.585534 4896 generic.go:334] "Generic (PLEG): container finished" podID="0e13f33c-acb4-421a-be84-4757ff7c3b74" containerID="71e9ad2c7f8e1409e222a5a29e82e3c4795fa2d58deab669795283992f6413fd" exitCode=0 Feb 18 15:09:17 crc kubenswrapper[4896]: I0218 15:09:17.585745 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/openstack-galera-2" event={"ID":"0e13f33c-acb4-421a-be84-4757ff7c3b74","Type":"ContainerDied","Data":"71e9ad2c7f8e1409e222a5a29e82e3c4795fa2d58deab669795283992f6413fd"} Feb 18 15:09:17 crc kubenswrapper[4896]: I0218 15:09:17.588032 4896 generic.go:334] "Generic (PLEG): container finished" podID="1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed" containerID="5cfc8cd25d326b00c6e48ecc4a150c223599feef13c2ab0e697363415ff36dd9" exitCode=0 Feb 18 15:09:17 crc kubenswrapper[4896]: I0218 15:09:17.588084 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/openstack-galera-0" event={"ID":"1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed","Type":"ContainerDied","Data":"5cfc8cd25d326b00c6e48ecc4a150c223599feef13c2ab0e697363415ff36dd9"} Feb 18 15:09:17 crc kubenswrapper[4896]: I0218 15:09:17.590014 4896 generic.go:334] "Generic (PLEG): container finished" podID="1a1927a6-afa4-41c7-95ad-22d7bba44d38" containerID="a284c2e2fe2036ab6c2b4010bee31b3a50cd6c4aeb0284a6147f2ad000d0e748" exitCode=0 Feb 18 15:09:17 crc kubenswrapper[4896]: I0218 15:09:17.590034 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/openstack-galera-1" event={"ID":"1a1927a6-afa4-41c7-95ad-22d7bba44d38","Type":"ContainerDied","Data":"a284c2e2fe2036ab6c2b4010bee31b3a50cd6c4aeb0284a6147f2ad000d0e748"} Feb 18 15:09:18 crc kubenswrapper[4896]: I0218 15:09:18.598446 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/openstack-galera-2" event={"ID":"0e13f33c-acb4-421a-be84-4757ff7c3b74","Type":"ContainerStarted","Data":"38a7ff893516ea0a43be5384b4f6baa16e39e2de5b0477bcd88633d946aa2c94"} Feb 18 15:09:18 crc kubenswrapper[4896]: I0218 15:09:18.601736 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/openstack-galera-0" event={"ID":"1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed","Type":"ContainerStarted","Data":"9cbb1e77fcd4a26fa8014fc2b9c94da35a845f243fbad63aa810052bcc845fd2"} Feb 18 15:09:18 crc kubenswrapper[4896]: I0218 15:09:18.603936 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/openstack-galera-1" event={"ID":"1a1927a6-afa4-41c7-95ad-22d7bba44d38","Type":"ContainerStarted","Data":"f09388b4f448f58d734175ad2b870801ed54e7772aeea252ef53863d05d51444"} Feb 18 15:09:18 crc kubenswrapper[4896]: I0218 15:09:18.634021 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="test-operator-kuttl-tests/openstack-galera-2" podStartSLOduration=7.787223469 podStartE2EDuration="19.634006408s" podCreationTimestamp="2026-02-18 15:08:59 +0000 UTC" firstStartedPulling="2026-02-18 15:09:01.293684558 +0000 UTC m=+840.223018781" lastFinishedPulling="2026-02-18 15:09:13.140467507 +0000 UTC m=+852.069801720" observedRunningTime="2026-02-18 15:09:18.616620587 +0000 UTC m=+857.545954800" watchObservedRunningTime="2026-02-18 15:09:18.634006408 +0000 UTC m=+857.563340611" Feb 18 15:09:18 crc kubenswrapper[4896]: I0218 15:09:18.638839 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="test-operator-kuttl-tests/openstack-galera-0" podStartSLOduration=7.985975226 podStartE2EDuration="19.638812935s" podCreationTimestamp="2026-02-18 15:08:59 +0000 UTC" firstStartedPulling="2026-02-18 15:09:01.437081239 +0000 UTC m=+840.366415452" lastFinishedPulling="2026-02-18 15:09:13.089918948 +0000 UTC m=+852.019253161" observedRunningTime="2026-02-18 15:09:18.633273658 +0000 UTC m=+857.562607871" watchObservedRunningTime="2026-02-18 15:09:18.638812935 +0000 UTC m=+857.568147148" Feb 18 15:09:18 crc kubenswrapper[4896]: I0218 15:09:18.654258 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="test-operator-kuttl-tests/openstack-galera-1" podStartSLOduration=7.712694667 podStartE2EDuration="19.654243344s" podCreationTimestamp="2026-02-18 15:08:59 +0000 UTC" firstStartedPulling="2026-02-18 15:09:01.180826641 +0000 UTC m=+840.110160854" lastFinishedPulling="2026-02-18 15:09:13.122375318 +0000 UTC m=+852.051709531" observedRunningTime="2026-02-18 15:09:18.651206083 +0000 UTC m=+857.580540296" watchObservedRunningTime="2026-02-18 15:09:18.654243344 +0000 UTC m=+857.583577557" Feb 18 15:09:20 crc kubenswrapper[4896]: I0218 15:09:20.700408 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="test-operator-kuttl-tests/openstack-galera-1" Feb 18 15:09:20 crc kubenswrapper[4896]: I0218 15:09:20.701513 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="test-operator-kuttl-tests/openstack-galera-1" Feb 18 15:09:20 crc kubenswrapper[4896]: I0218 15:09:20.965268 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="test-operator-kuttl-tests/openstack-galera-0" Feb 18 15:09:20 crc kubenswrapper[4896]: I0218 15:09:20.965314 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="test-operator-kuttl-tests/openstack-galera-0" Feb 18 15:09:20 crc kubenswrapper[4896]: I0218 15:09:20.983037 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="test-operator-kuttl-tests/openstack-galera-2" Feb 18 15:09:20 crc kubenswrapper[4896]: I0218 15:09:20.983283 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="test-operator-kuttl-tests/openstack-galera-2" Feb 18 15:09:21 crc kubenswrapper[4896]: I0218 15:09:21.098472 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-5d86b96dc9-dhl2n" Feb 18 15:09:22 crc kubenswrapper[4896]: I0218 15:09:22.009050 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["test-operator-kuttl-tests/memcached-0"] Feb 18 15:09:22 crc kubenswrapper[4896]: I0218 15:09:22.009853 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="test-operator-kuttl-tests/memcached-0" Feb 18 15:09:22 crc kubenswrapper[4896]: I0218 15:09:22.011760 4896 reflector.go:368] Caches populated for *v1.Secret from object-"test-operator-kuttl-tests"/"memcached-memcached-dockercfg-g9tqd" Feb 18 15:09:22 crc kubenswrapper[4896]: I0218 15:09:22.012098 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"test-operator-kuttl-tests"/"memcached-config-data" Feb 18 15:09:22 crc kubenswrapper[4896]: I0218 15:09:22.022691 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["test-operator-kuttl-tests/memcached-0"] Feb 18 15:09:22 crc kubenswrapper[4896]: I0218 15:09:22.153977 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxqnk\" (UniqueName: \"kubernetes.io/projected/7dfbd600-c900-4217-a46a-cb07e68791e2-kube-api-access-xxqnk\") pod \"memcached-0\" (UID: \"7dfbd600-c900-4217-a46a-cb07e68791e2\") " pod="test-operator-kuttl-tests/memcached-0" Feb 18 15:09:22 crc kubenswrapper[4896]: I0218 15:09:22.154300 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7dfbd600-c900-4217-a46a-cb07e68791e2-config-data\") pod \"memcached-0\" (UID: \"7dfbd600-c900-4217-a46a-cb07e68791e2\") " pod="test-operator-kuttl-tests/memcached-0" Feb 18 15:09:22 crc kubenswrapper[4896]: I0218 15:09:22.154331 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7dfbd600-c900-4217-a46a-cb07e68791e2-kolla-config\") pod \"memcached-0\" (UID: \"7dfbd600-c900-4217-a46a-cb07e68791e2\") " pod="test-operator-kuttl-tests/memcached-0" Feb 18 15:09:22 crc kubenswrapper[4896]: I0218 15:09:22.255652 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7dfbd600-c900-4217-a46a-cb07e68791e2-config-data\") pod \"memcached-0\" (UID: \"7dfbd600-c900-4217-a46a-cb07e68791e2\") " pod="test-operator-kuttl-tests/memcached-0" Feb 18 15:09:22 crc kubenswrapper[4896]: I0218 15:09:22.255725 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7dfbd600-c900-4217-a46a-cb07e68791e2-kolla-config\") pod \"memcached-0\" (UID: \"7dfbd600-c900-4217-a46a-cb07e68791e2\") " pod="test-operator-kuttl-tests/memcached-0" Feb 18 15:09:22 crc kubenswrapper[4896]: I0218 15:09:22.255793 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxqnk\" (UniqueName: \"kubernetes.io/projected/7dfbd600-c900-4217-a46a-cb07e68791e2-kube-api-access-xxqnk\") pod \"memcached-0\" (UID: \"7dfbd600-c900-4217-a46a-cb07e68791e2\") " pod="test-operator-kuttl-tests/memcached-0" Feb 18 15:09:22 crc kubenswrapper[4896]: I0218 15:09:22.256531 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/7dfbd600-c900-4217-a46a-cb07e68791e2-config-data\") pod \"memcached-0\" (UID: \"7dfbd600-c900-4217-a46a-cb07e68791e2\") " pod="test-operator-kuttl-tests/memcached-0" Feb 18 15:09:22 crc kubenswrapper[4896]: I0218 15:09:22.256612 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7dfbd600-c900-4217-a46a-cb07e68791e2-kolla-config\") pod \"memcached-0\" (UID: \"7dfbd600-c900-4217-a46a-cb07e68791e2\") " pod="test-operator-kuttl-tests/memcached-0" Feb 18 15:09:22 crc kubenswrapper[4896]: I0218 15:09:22.273790 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxqnk\" (UniqueName: \"kubernetes.io/projected/7dfbd600-c900-4217-a46a-cb07e68791e2-kube-api-access-xxqnk\") pod \"memcached-0\" (UID: \"7dfbd600-c900-4217-a46a-cb07e68791e2\") " pod="test-operator-kuttl-tests/memcached-0" Feb 18 15:09:22 crc kubenswrapper[4896]: I0218 15:09:22.278053 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-sn6hw"] Feb 18 15:09:22 crc kubenswrapper[4896]: I0218 15:09:22.278776 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-sn6hw" Feb 18 15:09:22 crc kubenswrapper[4896]: I0218 15:09:22.280400 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-index-dockercfg-pkjzh" Feb 18 15:09:22 crc kubenswrapper[4896]: I0218 15:09:22.307028 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-sn6hw"] Feb 18 15:09:22 crc kubenswrapper[4896]: I0218 15:09:22.339229 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="test-operator-kuttl-tests/memcached-0" Feb 18 15:09:22 crc kubenswrapper[4896]: I0218 15:09:22.356924 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcncq\" (UniqueName: \"kubernetes.io/projected/b8d068ab-922c-4ae5-99be-884fe154b953-kube-api-access-tcncq\") pod \"rabbitmq-cluster-operator-index-sn6hw\" (UID: \"b8d068ab-922c-4ae5-99be-884fe154b953\") " pod="openstack-operators/rabbitmq-cluster-operator-index-sn6hw" Feb 18 15:09:22 crc kubenswrapper[4896]: I0218 15:09:22.458360 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcncq\" (UniqueName: \"kubernetes.io/projected/b8d068ab-922c-4ae5-99be-884fe154b953-kube-api-access-tcncq\") pod \"rabbitmq-cluster-operator-index-sn6hw\" (UID: \"b8d068ab-922c-4ae5-99be-884fe154b953\") " pod="openstack-operators/rabbitmq-cluster-operator-index-sn6hw" Feb 18 15:09:22 crc kubenswrapper[4896]: I0218 15:09:22.477869 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcncq\" (UniqueName: \"kubernetes.io/projected/b8d068ab-922c-4ae5-99be-884fe154b953-kube-api-access-tcncq\") pod \"rabbitmq-cluster-operator-index-sn6hw\" (UID: \"b8d068ab-922c-4ae5-99be-884fe154b953\") " pod="openstack-operators/rabbitmq-cluster-operator-index-sn6hw" Feb 18 15:09:22 crc kubenswrapper[4896]: I0218 15:09:22.653052 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-sn6hw" Feb 18 15:09:22 crc kubenswrapper[4896]: I0218 15:09:22.756544 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["test-operator-kuttl-tests/memcached-0"] Feb 18 15:09:23 crc kubenswrapper[4896]: I0218 15:09:23.111798 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-sn6hw"] Feb 18 15:09:23 crc kubenswrapper[4896]: W0218 15:09:23.126432 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb8d068ab_922c_4ae5_99be_884fe154b953.slice/crio-6e45c838452dd614eb8337859e0fd0d52611c06eff9b65f21db425eb44f17e99 WatchSource:0}: Error finding container 6e45c838452dd614eb8337859e0fd0d52611c06eff9b65f21db425eb44f17e99: Status 404 returned error can't find the container with id 6e45c838452dd614eb8337859e0fd0d52611c06eff9b65f21db425eb44f17e99 Feb 18 15:09:23 crc kubenswrapper[4896]: I0218 15:09:23.631823 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/memcached-0" event={"ID":"7dfbd600-c900-4217-a46a-cb07e68791e2","Type":"ContainerStarted","Data":"e43b7c5d2a703963d0677136d47de4f75011964aa79fabcb494cf33efc7f163f"} Feb 18 15:09:23 crc kubenswrapper[4896]: I0218 15:09:23.632977 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-sn6hw" event={"ID":"b8d068ab-922c-4ae5-99be-884fe154b953","Type":"ContainerStarted","Data":"6e45c838452dd614eb8337859e0fd0d52611c06eff9b65f21db425eb44f17e99"} Feb 18 15:09:24 crc kubenswrapper[4896]: I0218 15:09:24.638587 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/memcached-0" event={"ID":"7dfbd600-c900-4217-a46a-cb07e68791e2","Type":"ContainerStarted","Data":"8119838908c7b61deb466b659828670620a508932994ff99453d18aafc7d3e88"} Feb 18 15:09:24 crc kubenswrapper[4896]: I0218 15:09:24.646691 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="test-operator-kuttl-tests/memcached-0" Feb 18 15:09:24 crc kubenswrapper[4896]: I0218 15:09:24.673009 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="test-operator-kuttl-tests/memcached-0" podStartSLOduration=1.9599661419999999 podStartE2EDuration="3.672985627s" podCreationTimestamp="2026-02-18 15:09:21 +0000 UTC" firstStartedPulling="2026-02-18 15:09:22.762832841 +0000 UTC m=+861.692167044" lastFinishedPulling="2026-02-18 15:09:24.475852316 +0000 UTC m=+863.405186529" observedRunningTime="2026-02-18 15:09:24.666400203 +0000 UTC m=+863.595734426" watchObservedRunningTime="2026-02-18 15:09:24.672985627 +0000 UTC m=+863.602319840" Feb 18 15:09:25 crc kubenswrapper[4896]: E0218 15:09:25.787407 4896 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.156:45994->38.102.83.156:39147: write tcp 38.102.83.156:45994->38.102.83.156:39147: write: broken pipe Feb 18 15:09:26 crc kubenswrapper[4896]: I0218 15:09:26.471568 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-sn6hw"] Feb 18 15:09:27 crc kubenswrapper[4896]: I0218 15:09:27.079947 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-vtpwb"] Feb 18 15:09:27 crc kubenswrapper[4896]: I0218 15:09:27.081109 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-vtpwb" Feb 18 15:09:27 crc kubenswrapper[4896]: I0218 15:09:27.092939 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-vtpwb"] Feb 18 15:09:27 crc kubenswrapper[4896]: I0218 15:09:27.162680 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="test-operator-kuttl-tests/openstack-galera-2" Feb 18 15:09:27 crc kubenswrapper[4896]: I0218 15:09:27.226255 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="test-operator-kuttl-tests/openstack-galera-2" Feb 18 15:09:27 crc kubenswrapper[4896]: I0218 15:09:27.232523 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p88mb\" (UniqueName: \"kubernetes.io/projected/b56a5e00-18b9-42a9-9c39-a9a70fe7ab30-kube-api-access-p88mb\") pod \"rabbitmq-cluster-operator-index-vtpwb\" (UID: \"b56a5e00-18b9-42a9-9c39-a9a70fe7ab30\") " pod="openstack-operators/rabbitmq-cluster-operator-index-vtpwb" Feb 18 15:09:27 crc kubenswrapper[4896]: I0218 15:09:27.334060 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p88mb\" (UniqueName: \"kubernetes.io/projected/b56a5e00-18b9-42a9-9c39-a9a70fe7ab30-kube-api-access-p88mb\") pod \"rabbitmq-cluster-operator-index-vtpwb\" (UID: \"b56a5e00-18b9-42a9-9c39-a9a70fe7ab30\") " pod="openstack-operators/rabbitmq-cluster-operator-index-vtpwb" Feb 18 15:09:27 crc kubenswrapper[4896]: I0218 15:09:27.357901 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p88mb\" (UniqueName: \"kubernetes.io/projected/b56a5e00-18b9-42a9-9c39-a9a70fe7ab30-kube-api-access-p88mb\") pod \"rabbitmq-cluster-operator-index-vtpwb\" (UID: \"b56a5e00-18b9-42a9-9c39-a9a70fe7ab30\") " pod="openstack-operators/rabbitmq-cluster-operator-index-vtpwb" Feb 18 15:09:27 crc kubenswrapper[4896]: I0218 15:09:27.440529 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-vtpwb" Feb 18 15:09:27 crc kubenswrapper[4896]: I0218 15:09:27.693364 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-sn6hw" event={"ID":"b8d068ab-922c-4ae5-99be-884fe154b953","Type":"ContainerStarted","Data":"fdbec2b68078d86946317a4e43badda20e4405fbd26abce8898bcf69683c6996"} Feb 18 15:09:27 crc kubenswrapper[4896]: I0218 15:09:27.693395 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/rabbitmq-cluster-operator-index-sn6hw" podUID="b8d068ab-922c-4ae5-99be-884fe154b953" containerName="registry-server" containerID="cri-o://fdbec2b68078d86946317a4e43badda20e4405fbd26abce8898bcf69683c6996" gracePeriod=2 Feb 18 15:09:27 crc kubenswrapper[4896]: I0218 15:09:27.713512 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-sn6hw" podStartSLOduration=2.092181424 podStartE2EDuration="5.713496754s" podCreationTimestamp="2026-02-18 15:09:22 +0000 UTC" firstStartedPulling="2026-02-18 15:09:23.127872201 +0000 UTC m=+862.057206414" lastFinishedPulling="2026-02-18 15:09:26.749187531 +0000 UTC m=+865.678521744" observedRunningTime="2026-02-18 15:09:27.70846512 +0000 UTC m=+866.637799343" watchObservedRunningTime="2026-02-18 15:09:27.713496754 +0000 UTC m=+866.642830957" Feb 18 15:09:27 crc kubenswrapper[4896]: I0218 15:09:27.844408 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-vtpwb"] Feb 18 15:09:27 crc kubenswrapper[4896]: W0218 15:09:27.853061 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb56a5e00_18b9_42a9_9c39_a9a70fe7ab30.slice/crio-b523226bee032947daa6571f501641932b7b0a68f5eed91a99d2680aeee80388 WatchSource:0}: Error finding container b523226bee032947daa6571f501641932b7b0a68f5eed91a99d2680aeee80388: Status 404 returned error can't find the container with id b523226bee032947daa6571f501641932b7b0a68f5eed91a99d2680aeee80388 Feb 18 15:09:28 crc kubenswrapper[4896]: I0218 15:09:28.110549 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-sn6hw" Feb 18 15:09:28 crc kubenswrapper[4896]: I0218 15:09:28.246355 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tcncq\" (UniqueName: \"kubernetes.io/projected/b8d068ab-922c-4ae5-99be-884fe154b953-kube-api-access-tcncq\") pod \"b8d068ab-922c-4ae5-99be-884fe154b953\" (UID: \"b8d068ab-922c-4ae5-99be-884fe154b953\") " Feb 18 15:09:28 crc kubenswrapper[4896]: I0218 15:09:28.255427 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8d068ab-922c-4ae5-99be-884fe154b953-kube-api-access-tcncq" (OuterVolumeSpecName: "kube-api-access-tcncq") pod "b8d068ab-922c-4ae5-99be-884fe154b953" (UID: "b8d068ab-922c-4ae5-99be-884fe154b953"). InnerVolumeSpecName "kube-api-access-tcncq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:09:28 crc kubenswrapper[4896]: I0218 15:09:28.347483 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tcncq\" (UniqueName: \"kubernetes.io/projected/b8d068ab-922c-4ae5-99be-884fe154b953-kube-api-access-tcncq\") on node \"crc\" DevicePath \"\"" Feb 18 15:09:28 crc kubenswrapper[4896]: I0218 15:09:28.700527 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-vtpwb" event={"ID":"b56a5e00-18b9-42a9-9c39-a9a70fe7ab30","Type":"ContainerStarted","Data":"7a18dc88b858a7ec96559b8aba151018e78d69ea489cb8498ae7fcb71b6c2aab"} Feb 18 15:09:28 crc kubenswrapper[4896]: I0218 15:09:28.700857 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-vtpwb" event={"ID":"b56a5e00-18b9-42a9-9c39-a9a70fe7ab30","Type":"ContainerStarted","Data":"b523226bee032947daa6571f501641932b7b0a68f5eed91a99d2680aeee80388"} Feb 18 15:09:28 crc kubenswrapper[4896]: I0218 15:09:28.702103 4896 generic.go:334] "Generic (PLEG): container finished" podID="b8d068ab-922c-4ae5-99be-884fe154b953" containerID="fdbec2b68078d86946317a4e43badda20e4405fbd26abce8898bcf69683c6996" exitCode=0 Feb 18 15:09:28 crc kubenswrapper[4896]: I0218 15:09:28.702140 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-sn6hw" Feb 18 15:09:28 crc kubenswrapper[4896]: I0218 15:09:28.702143 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-sn6hw" event={"ID":"b8d068ab-922c-4ae5-99be-884fe154b953","Type":"ContainerDied","Data":"fdbec2b68078d86946317a4e43badda20e4405fbd26abce8898bcf69683c6996"} Feb 18 15:09:28 crc kubenswrapper[4896]: I0218 15:09:28.702306 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-sn6hw" event={"ID":"b8d068ab-922c-4ae5-99be-884fe154b953","Type":"ContainerDied","Data":"6e45c838452dd614eb8337859e0fd0d52611c06eff9b65f21db425eb44f17e99"} Feb 18 15:09:28 crc kubenswrapper[4896]: I0218 15:09:28.702337 4896 scope.go:117] "RemoveContainer" containerID="fdbec2b68078d86946317a4e43badda20e4405fbd26abce8898bcf69683c6996" Feb 18 15:09:28 crc kubenswrapper[4896]: I0218 15:09:28.720083 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-vtpwb" podStartSLOduration=1.3090442869999999 podStartE2EDuration="1.720063795s" podCreationTimestamp="2026-02-18 15:09:27 +0000 UTC" firstStartedPulling="2026-02-18 15:09:27.857446956 +0000 UTC m=+866.786781159" lastFinishedPulling="2026-02-18 15:09:28.268466454 +0000 UTC m=+867.197800667" observedRunningTime="2026-02-18 15:09:28.712950246 +0000 UTC m=+867.642284459" watchObservedRunningTime="2026-02-18 15:09:28.720063795 +0000 UTC m=+867.649398018" Feb 18 15:09:28 crc kubenswrapper[4896]: I0218 15:09:28.720646 4896 scope.go:117] "RemoveContainer" containerID="fdbec2b68078d86946317a4e43badda20e4405fbd26abce8898bcf69683c6996" Feb 18 15:09:28 crc kubenswrapper[4896]: E0218 15:09:28.721018 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdbec2b68078d86946317a4e43badda20e4405fbd26abce8898bcf69683c6996\": container with ID starting with fdbec2b68078d86946317a4e43badda20e4405fbd26abce8898bcf69683c6996 not found: ID does not exist" containerID="fdbec2b68078d86946317a4e43badda20e4405fbd26abce8898bcf69683c6996" Feb 18 15:09:28 crc kubenswrapper[4896]: I0218 15:09:28.721065 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdbec2b68078d86946317a4e43badda20e4405fbd26abce8898bcf69683c6996"} err="failed to get container status \"fdbec2b68078d86946317a4e43badda20e4405fbd26abce8898bcf69683c6996\": rpc error: code = NotFound desc = could not find container \"fdbec2b68078d86946317a4e43badda20e4405fbd26abce8898bcf69683c6996\": container with ID starting with fdbec2b68078d86946317a4e43badda20e4405fbd26abce8898bcf69683c6996 not found: ID does not exist" Feb 18 15:09:28 crc kubenswrapper[4896]: I0218 15:09:28.729629 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-sn6hw"] Feb 18 15:09:28 crc kubenswrapper[4896]: I0218 15:09:28.734821 4896 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-sn6hw"] Feb 18 15:09:29 crc kubenswrapper[4896]: I0218 15:09:29.426488 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["test-operator-kuttl-tests/root-account-create-update-45cql"] Feb 18 15:09:29 crc kubenswrapper[4896]: E0218 15:09:29.426766 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8d068ab-922c-4ae5-99be-884fe154b953" containerName="registry-server" Feb 18 15:09:29 crc kubenswrapper[4896]: I0218 15:09:29.426781 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8d068ab-922c-4ae5-99be-884fe154b953" containerName="registry-server" Feb 18 15:09:29 crc kubenswrapper[4896]: I0218 15:09:29.426917 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8d068ab-922c-4ae5-99be-884fe154b953" containerName="registry-server" Feb 18 15:09:29 crc kubenswrapper[4896]: I0218 15:09:29.427434 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="test-operator-kuttl-tests/root-account-create-update-45cql" Feb 18 15:09:29 crc kubenswrapper[4896]: I0218 15:09:29.429453 4896 reflector.go:368] Caches populated for *v1.Secret from object-"test-operator-kuttl-tests"/"openstack-mariadb-root-db-secret" Feb 18 15:09:29 crc kubenswrapper[4896]: I0218 15:09:29.473625 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["test-operator-kuttl-tests/root-account-create-update-45cql"] Feb 18 15:09:29 crc kubenswrapper[4896]: E0218 15:09:29.533430 4896 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.156:46068->38.102.83.156:39147: write tcp 38.102.83.156:46068->38.102.83.156:39147: write: broken pipe Feb 18 15:09:29 crc kubenswrapper[4896]: I0218 15:09:29.562011 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqlgg\" (UniqueName: \"kubernetes.io/projected/fe758f2a-b576-4713-9803-4d34178cfebf-kube-api-access-gqlgg\") pod \"root-account-create-update-45cql\" (UID: \"fe758f2a-b576-4713-9803-4d34178cfebf\") " pod="test-operator-kuttl-tests/root-account-create-update-45cql" Feb 18 15:09:29 crc kubenswrapper[4896]: I0218 15:09:29.562193 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe758f2a-b576-4713-9803-4d34178cfebf-operator-scripts\") pod \"root-account-create-update-45cql\" (UID: \"fe758f2a-b576-4713-9803-4d34178cfebf\") " pod="test-operator-kuttl-tests/root-account-create-update-45cql" Feb 18 15:09:29 crc kubenswrapper[4896]: I0218 15:09:29.663257 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqlgg\" (UniqueName: \"kubernetes.io/projected/fe758f2a-b576-4713-9803-4d34178cfebf-kube-api-access-gqlgg\") pod \"root-account-create-update-45cql\" (UID: \"fe758f2a-b576-4713-9803-4d34178cfebf\") " pod="test-operator-kuttl-tests/root-account-create-update-45cql" Feb 18 15:09:29 crc kubenswrapper[4896]: I0218 15:09:29.663349 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe758f2a-b576-4713-9803-4d34178cfebf-operator-scripts\") pod \"root-account-create-update-45cql\" (UID: \"fe758f2a-b576-4713-9803-4d34178cfebf\") " pod="test-operator-kuttl-tests/root-account-create-update-45cql" Feb 18 15:09:29 crc kubenswrapper[4896]: I0218 15:09:29.664099 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe758f2a-b576-4713-9803-4d34178cfebf-operator-scripts\") pod \"root-account-create-update-45cql\" (UID: \"fe758f2a-b576-4713-9803-4d34178cfebf\") " pod="test-operator-kuttl-tests/root-account-create-update-45cql" Feb 18 15:09:29 crc kubenswrapper[4896]: I0218 15:09:29.692095 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqlgg\" (UniqueName: \"kubernetes.io/projected/fe758f2a-b576-4713-9803-4d34178cfebf-kube-api-access-gqlgg\") pod \"root-account-create-update-45cql\" (UID: \"fe758f2a-b576-4713-9803-4d34178cfebf\") " pod="test-operator-kuttl-tests/root-account-create-update-45cql" Feb 18 15:09:29 crc kubenswrapper[4896]: I0218 15:09:29.746265 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="test-operator-kuttl-tests/root-account-create-update-45cql" Feb 18 15:09:29 crc kubenswrapper[4896]: I0218 15:09:29.936724 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8d068ab-922c-4ae5-99be-884fe154b953" path="/var/lib/kubelet/pods/b8d068ab-922c-4ae5-99be-884fe154b953/volumes" Feb 18 15:09:30 crc kubenswrapper[4896]: I0218 15:09:30.191033 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["test-operator-kuttl-tests/root-account-create-update-45cql"] Feb 18 15:09:30 crc kubenswrapper[4896]: W0218 15:09:30.191891 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe758f2a_b576_4713_9803_4d34178cfebf.slice/crio-bb2f78ca11c3f7cb400c7391b23fe0d00f5ea54e7a01f71b438b283d2bc3b9dd WatchSource:0}: Error finding container bb2f78ca11c3f7cb400c7391b23fe0d00f5ea54e7a01f71b438b283d2bc3b9dd: Status 404 returned error can't find the container with id bb2f78ca11c3f7cb400c7391b23fe0d00f5ea54e7a01f71b438b283d2bc3b9dd Feb 18 15:09:30 crc kubenswrapper[4896]: I0218 15:09:30.715183 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/root-account-create-update-45cql" event={"ID":"fe758f2a-b576-4713-9803-4d34178cfebf","Type":"ContainerStarted","Data":"f3d17a83509b51dad78c25b94c855ae20d8fb865435518f3a227dd89df01c39c"} Feb 18 15:09:30 crc kubenswrapper[4896]: I0218 15:09:30.715252 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/root-account-create-update-45cql" event={"ID":"fe758f2a-b576-4713-9803-4d34178cfebf","Type":"ContainerStarted","Data":"bb2f78ca11c3f7cb400c7391b23fe0d00f5ea54e7a01f71b438b283d2bc3b9dd"} Feb 18 15:09:30 crc kubenswrapper[4896]: I0218 15:09:30.733332 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="test-operator-kuttl-tests/root-account-create-update-45cql" podStartSLOduration=1.733314432 podStartE2EDuration="1.733314432s" podCreationTimestamp="2026-02-18 15:09:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:09:30.732752577 +0000 UTC m=+869.662086790" watchObservedRunningTime="2026-02-18 15:09:30.733314432 +0000 UTC m=+869.662648645" Feb 18 15:09:31 crc kubenswrapper[4896]: I0218 15:09:31.095760 4896 prober.go:107] "Probe failed" probeType="Readiness" pod="test-operator-kuttl-tests/openstack-galera-2" podUID="0e13f33c-acb4-421a-be84-4757ff7c3b74" containerName="galera" probeResult="failure" output=< Feb 18 15:09:31 crc kubenswrapper[4896]: wsrep_local_state_comment (Donor/Desynced) differs from Synced Feb 18 15:09:31 crc kubenswrapper[4896]: > Feb 18 15:09:31 crc kubenswrapper[4896]: I0218 15:09:31.722941 4896 generic.go:334] "Generic (PLEG): container finished" podID="fe758f2a-b576-4713-9803-4d34178cfebf" containerID="f3d17a83509b51dad78c25b94c855ae20d8fb865435518f3a227dd89df01c39c" exitCode=0 Feb 18 15:09:31 crc kubenswrapper[4896]: I0218 15:09:31.722983 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/root-account-create-update-45cql" event={"ID":"fe758f2a-b576-4713-9803-4d34178cfebf","Type":"ContainerDied","Data":"f3d17a83509b51dad78c25b94c855ae20d8fb865435518f3a227dd89df01c39c"} Feb 18 15:09:32 crc kubenswrapper[4896]: I0218 15:09:32.340408 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="test-operator-kuttl-tests/memcached-0" Feb 18 15:09:33 crc kubenswrapper[4896]: I0218 15:09:33.098624 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="test-operator-kuttl-tests/root-account-create-update-45cql" Feb 18 15:09:33 crc kubenswrapper[4896]: I0218 15:09:33.212875 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqlgg\" (UniqueName: \"kubernetes.io/projected/fe758f2a-b576-4713-9803-4d34178cfebf-kube-api-access-gqlgg\") pod \"fe758f2a-b576-4713-9803-4d34178cfebf\" (UID: \"fe758f2a-b576-4713-9803-4d34178cfebf\") " Feb 18 15:09:33 crc kubenswrapper[4896]: I0218 15:09:33.213025 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe758f2a-b576-4713-9803-4d34178cfebf-operator-scripts\") pod \"fe758f2a-b576-4713-9803-4d34178cfebf\" (UID: \"fe758f2a-b576-4713-9803-4d34178cfebf\") " Feb 18 15:09:33 crc kubenswrapper[4896]: I0218 15:09:33.213540 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe758f2a-b576-4713-9803-4d34178cfebf-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fe758f2a-b576-4713-9803-4d34178cfebf" (UID: "fe758f2a-b576-4713-9803-4d34178cfebf"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:09:33 crc kubenswrapper[4896]: I0218 15:09:33.222067 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe758f2a-b576-4713-9803-4d34178cfebf-kube-api-access-gqlgg" (OuterVolumeSpecName: "kube-api-access-gqlgg") pod "fe758f2a-b576-4713-9803-4d34178cfebf" (UID: "fe758f2a-b576-4713-9803-4d34178cfebf"). InnerVolumeSpecName "kube-api-access-gqlgg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:09:33 crc kubenswrapper[4896]: I0218 15:09:33.314434 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqlgg\" (UniqueName: \"kubernetes.io/projected/fe758f2a-b576-4713-9803-4d34178cfebf-kube-api-access-gqlgg\") on node \"crc\" DevicePath \"\"" Feb 18 15:09:33 crc kubenswrapper[4896]: I0218 15:09:33.314690 4896 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fe758f2a-b576-4713-9803-4d34178cfebf-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 15:09:33 crc kubenswrapper[4896]: I0218 15:09:33.734550 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/root-account-create-update-45cql" event={"ID":"fe758f2a-b576-4713-9803-4d34178cfebf","Type":"ContainerDied","Data":"bb2f78ca11c3f7cb400c7391b23fe0d00f5ea54e7a01f71b438b283d2bc3b9dd"} Feb 18 15:09:33 crc kubenswrapper[4896]: I0218 15:09:33.734590 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="test-operator-kuttl-tests/root-account-create-update-45cql" Feb 18 15:09:33 crc kubenswrapper[4896]: I0218 15:09:33.734596 4896 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb2f78ca11c3f7cb400c7391b23fe0d00f5ea54e7a01f71b438b283d2bc3b9dd" Feb 18 15:09:35 crc kubenswrapper[4896]: I0218 15:09:35.479395 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-q2wgt"] Feb 18 15:09:35 crc kubenswrapper[4896]: E0218 15:09:35.479623 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe758f2a-b576-4713-9803-4d34178cfebf" containerName="mariadb-account-create-update" Feb 18 15:09:35 crc kubenswrapper[4896]: I0218 15:09:35.479634 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe758f2a-b576-4713-9803-4d34178cfebf" containerName="mariadb-account-create-update" Feb 18 15:09:35 crc kubenswrapper[4896]: I0218 15:09:35.479747 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe758f2a-b576-4713-9803-4d34178cfebf" containerName="mariadb-account-create-update" Feb 18 15:09:35 crc kubenswrapper[4896]: I0218 15:09:35.480507 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q2wgt" Feb 18 15:09:35 crc kubenswrapper[4896]: I0218 15:09:35.543123 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b1335f8-9329-48fa-b924-f6729c81100d-utilities\") pod \"redhat-marketplace-q2wgt\" (UID: \"3b1335f8-9329-48fa-b924-f6729c81100d\") " pod="openshift-marketplace/redhat-marketplace-q2wgt" Feb 18 15:09:35 crc kubenswrapper[4896]: I0218 15:09:35.543475 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b1335f8-9329-48fa-b924-f6729c81100d-catalog-content\") pod \"redhat-marketplace-q2wgt\" (UID: \"3b1335f8-9329-48fa-b924-f6729c81100d\") " pod="openshift-marketplace/redhat-marketplace-q2wgt" Feb 18 15:09:35 crc kubenswrapper[4896]: I0218 15:09:35.543536 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x42dz\" (UniqueName: \"kubernetes.io/projected/3b1335f8-9329-48fa-b924-f6729c81100d-kube-api-access-x42dz\") pod \"redhat-marketplace-q2wgt\" (UID: \"3b1335f8-9329-48fa-b924-f6729c81100d\") " pod="openshift-marketplace/redhat-marketplace-q2wgt" Feb 18 15:09:35 crc kubenswrapper[4896]: I0218 15:09:35.584408 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q2wgt"] Feb 18 15:09:35 crc kubenswrapper[4896]: I0218 15:09:35.644837 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b1335f8-9329-48fa-b924-f6729c81100d-utilities\") pod \"redhat-marketplace-q2wgt\" (UID: \"3b1335f8-9329-48fa-b924-f6729c81100d\") " pod="openshift-marketplace/redhat-marketplace-q2wgt" Feb 18 15:09:35 crc kubenswrapper[4896]: I0218 15:09:35.644892 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b1335f8-9329-48fa-b924-f6729c81100d-catalog-content\") pod \"redhat-marketplace-q2wgt\" (UID: \"3b1335f8-9329-48fa-b924-f6729c81100d\") " pod="openshift-marketplace/redhat-marketplace-q2wgt" Feb 18 15:09:35 crc kubenswrapper[4896]: I0218 15:09:35.644952 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x42dz\" (UniqueName: \"kubernetes.io/projected/3b1335f8-9329-48fa-b924-f6729c81100d-kube-api-access-x42dz\") pod \"redhat-marketplace-q2wgt\" (UID: \"3b1335f8-9329-48fa-b924-f6729c81100d\") " pod="openshift-marketplace/redhat-marketplace-q2wgt" Feb 18 15:09:35 crc kubenswrapper[4896]: I0218 15:09:35.645644 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b1335f8-9329-48fa-b924-f6729c81100d-catalog-content\") pod \"redhat-marketplace-q2wgt\" (UID: \"3b1335f8-9329-48fa-b924-f6729c81100d\") " pod="openshift-marketplace/redhat-marketplace-q2wgt" Feb 18 15:09:35 crc kubenswrapper[4896]: I0218 15:09:35.645648 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b1335f8-9329-48fa-b924-f6729c81100d-utilities\") pod \"redhat-marketplace-q2wgt\" (UID: \"3b1335f8-9329-48fa-b924-f6729c81100d\") " pod="openshift-marketplace/redhat-marketplace-q2wgt" Feb 18 15:09:35 crc kubenswrapper[4896]: I0218 15:09:35.664072 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x42dz\" (UniqueName: \"kubernetes.io/projected/3b1335f8-9329-48fa-b924-f6729c81100d-kube-api-access-x42dz\") pod \"redhat-marketplace-q2wgt\" (UID: \"3b1335f8-9329-48fa-b924-f6729c81100d\") " pod="openshift-marketplace/redhat-marketplace-q2wgt" Feb 18 15:09:35 crc kubenswrapper[4896]: I0218 15:09:35.841778 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q2wgt" Feb 18 15:09:36 crc kubenswrapper[4896]: I0218 15:09:36.289818 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q2wgt"] Feb 18 15:09:36 crc kubenswrapper[4896]: W0218 15:09:36.295872 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b1335f8_9329_48fa_b924_f6729c81100d.slice/crio-551e7d2d0fc16f626994bbf110f94ffe8390851a7ad4ba85f5ac5e23f557421c WatchSource:0}: Error finding container 551e7d2d0fc16f626994bbf110f94ffe8390851a7ad4ba85f5ac5e23f557421c: Status 404 returned error can't find the container with id 551e7d2d0fc16f626994bbf110f94ffe8390851a7ad4ba85f5ac5e23f557421c Feb 18 15:09:36 crc kubenswrapper[4896]: I0218 15:09:36.767061 4896 generic.go:334] "Generic (PLEG): container finished" podID="3b1335f8-9329-48fa-b924-f6729c81100d" containerID="674325e530bd8a879c64f8a1728ee47d9044472076db2451ff2017c549f26f1a" exitCode=0 Feb 18 15:09:36 crc kubenswrapper[4896]: I0218 15:09:36.767108 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q2wgt" event={"ID":"3b1335f8-9329-48fa-b924-f6729c81100d","Type":"ContainerDied","Data":"674325e530bd8a879c64f8a1728ee47d9044472076db2451ff2017c549f26f1a"} Feb 18 15:09:36 crc kubenswrapper[4896]: I0218 15:09:36.767334 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q2wgt" event={"ID":"3b1335f8-9329-48fa-b924-f6729c81100d","Type":"ContainerStarted","Data":"551e7d2d0fc16f626994bbf110f94ffe8390851a7ad4ba85f5ac5e23f557421c"} Feb 18 15:09:37 crc kubenswrapper[4896]: I0218 15:09:37.440774 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/rabbitmq-cluster-operator-index-vtpwb" Feb 18 15:09:37 crc kubenswrapper[4896]: I0218 15:09:37.440850 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/rabbitmq-cluster-operator-index-vtpwb" Feb 18 15:09:37 crc kubenswrapper[4896]: I0218 15:09:37.482887 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/rabbitmq-cluster-operator-index-vtpwb" Feb 18 15:09:37 crc kubenswrapper[4896]: I0218 15:09:37.774304 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q2wgt" event={"ID":"3b1335f8-9329-48fa-b924-f6729c81100d","Type":"ContainerStarted","Data":"b8fcbaade3af77067e9703169f6034047980a492e605b2b8bc7213849859945b"} Feb 18 15:09:37 crc kubenswrapper[4896]: I0218 15:09:37.818030 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/rabbitmq-cluster-operator-index-vtpwb" Feb 18 15:09:38 crc kubenswrapper[4896]: I0218 15:09:38.764602 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="test-operator-kuttl-tests/openstack-galera-1" Feb 18 15:09:38 crc kubenswrapper[4896]: I0218 15:09:38.788070 4896 generic.go:334] "Generic (PLEG): container finished" podID="3b1335f8-9329-48fa-b924-f6729c81100d" containerID="b8fcbaade3af77067e9703169f6034047980a492e605b2b8bc7213849859945b" exitCode=0 Feb 18 15:09:38 crc kubenswrapper[4896]: I0218 15:09:38.788255 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q2wgt" event={"ID":"3b1335f8-9329-48fa-b924-f6729c81100d","Type":"ContainerDied","Data":"b8fcbaade3af77067e9703169f6034047980a492e605b2b8bc7213849859945b"} Feb 18 15:09:38 crc kubenswrapper[4896]: I0218 15:09:38.844859 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="test-operator-kuttl-tests/openstack-galera-1" Feb 18 15:09:39 crc kubenswrapper[4896]: I0218 15:09:39.798348 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q2wgt" event={"ID":"3b1335f8-9329-48fa-b924-f6729c81100d","Type":"ContainerStarted","Data":"f370806378c462f1b8f1109c1095a8e944e4dc74ad68fe67da21be95e62c0ab7"} Feb 18 15:09:39 crc kubenswrapper[4896]: I0218 15:09:39.817696 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-q2wgt" podStartSLOduration=2.179852536 podStartE2EDuration="4.817674335s" podCreationTimestamp="2026-02-18 15:09:35 +0000 UTC" firstStartedPulling="2026-02-18 15:09:36.768314335 +0000 UTC m=+875.697648548" lastFinishedPulling="2026-02-18 15:09:39.406136144 +0000 UTC m=+878.335470347" observedRunningTime="2026-02-18 15:09:39.813660469 +0000 UTC m=+878.742994702" watchObservedRunningTime="2026-02-18 15:09:39.817674335 +0000 UTC m=+878.747008548" Feb 18 15:09:41 crc kubenswrapper[4896]: I0218 15:09:41.778892 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="test-operator-kuttl-tests/openstack-galera-0" Feb 18 15:09:41 crc kubenswrapper[4896]: I0218 15:09:41.846713 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="test-operator-kuttl-tests/openstack-galera-0" Feb 18 15:09:44 crc kubenswrapper[4896]: I0218 15:09:44.516736 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp"] Feb 18 15:09:44 crc kubenswrapper[4896]: I0218 15:09:44.518136 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp" Feb 18 15:09:44 crc kubenswrapper[4896]: I0218 15:09:44.521160 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-9lxwl" Feb 18 15:09:44 crc kubenswrapper[4896]: I0218 15:09:44.533331 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp"] Feb 18 15:09:44 crc kubenswrapper[4896]: I0218 15:09:44.563416 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpjj4\" (UniqueName: \"kubernetes.io/projected/c6747e4d-4e79-4199-942b-8354cbafa962-kube-api-access-bpjj4\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp\" (UID: \"c6747e4d-4e79-4199-942b-8354cbafa962\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp" Feb 18 15:09:44 crc kubenswrapper[4896]: I0218 15:09:44.563503 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c6747e4d-4e79-4199-942b-8354cbafa962-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp\" (UID: \"c6747e4d-4e79-4199-942b-8354cbafa962\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp" Feb 18 15:09:44 crc kubenswrapper[4896]: I0218 15:09:44.563528 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c6747e4d-4e79-4199-942b-8354cbafa962-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp\" (UID: \"c6747e4d-4e79-4199-942b-8354cbafa962\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp" Feb 18 15:09:44 crc kubenswrapper[4896]: I0218 15:09:44.665061 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpjj4\" (UniqueName: \"kubernetes.io/projected/c6747e4d-4e79-4199-942b-8354cbafa962-kube-api-access-bpjj4\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp\" (UID: \"c6747e4d-4e79-4199-942b-8354cbafa962\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp" Feb 18 15:09:44 crc kubenswrapper[4896]: I0218 15:09:44.665176 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c6747e4d-4e79-4199-942b-8354cbafa962-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp\" (UID: \"c6747e4d-4e79-4199-942b-8354cbafa962\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp" Feb 18 15:09:44 crc kubenswrapper[4896]: I0218 15:09:44.665212 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c6747e4d-4e79-4199-942b-8354cbafa962-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp\" (UID: \"c6747e4d-4e79-4199-942b-8354cbafa962\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp" Feb 18 15:09:44 crc kubenswrapper[4896]: I0218 15:09:44.665733 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c6747e4d-4e79-4199-942b-8354cbafa962-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp\" (UID: \"c6747e4d-4e79-4199-942b-8354cbafa962\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp" Feb 18 15:09:44 crc kubenswrapper[4896]: I0218 15:09:44.665779 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c6747e4d-4e79-4199-942b-8354cbafa962-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp\" (UID: \"c6747e4d-4e79-4199-942b-8354cbafa962\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp" Feb 18 15:09:44 crc kubenswrapper[4896]: I0218 15:09:44.683166 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpjj4\" (UniqueName: \"kubernetes.io/projected/c6747e4d-4e79-4199-942b-8354cbafa962-kube-api-access-bpjj4\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp\" (UID: \"c6747e4d-4e79-4199-942b-8354cbafa962\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp" Feb 18 15:09:44 crc kubenswrapper[4896]: I0218 15:09:44.835812 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp" Feb 18 15:09:45 crc kubenswrapper[4896]: I0218 15:09:45.227300 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp"] Feb 18 15:09:45 crc kubenswrapper[4896]: I0218 15:09:45.834614 4896 generic.go:334] "Generic (PLEG): container finished" podID="c6747e4d-4e79-4199-942b-8354cbafa962" containerID="7f582534c69e1f9cf8efd68411ee18052dfd07fde35269824a7b751ea401ab38" exitCode=0 Feb 18 15:09:45 crc kubenswrapper[4896]: I0218 15:09:45.834653 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp" event={"ID":"c6747e4d-4e79-4199-942b-8354cbafa962","Type":"ContainerDied","Data":"7f582534c69e1f9cf8efd68411ee18052dfd07fde35269824a7b751ea401ab38"} Feb 18 15:09:45 crc kubenswrapper[4896]: I0218 15:09:45.834677 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp" event={"ID":"c6747e4d-4e79-4199-942b-8354cbafa962","Type":"ContainerStarted","Data":"09751550d2bce355911ac793f754482942f6dc6cddc451bff768d592dddb2b9b"} Feb 18 15:09:45 crc kubenswrapper[4896]: I0218 15:09:45.842511 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-q2wgt" Feb 18 15:09:45 crc kubenswrapper[4896]: I0218 15:09:45.842699 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-q2wgt" Feb 18 15:09:45 crc kubenswrapper[4896]: I0218 15:09:45.881035 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-q2wgt" Feb 18 15:09:46 crc kubenswrapper[4896]: I0218 15:09:46.841776 4896 generic.go:334] "Generic (PLEG): container finished" podID="c6747e4d-4e79-4199-942b-8354cbafa962" containerID="a1b71887a857052978d270a767bb6122b203861b5ff1580a88cf508f98bcc19b" exitCode=0 Feb 18 15:09:46 crc kubenswrapper[4896]: I0218 15:09:46.841809 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp" event={"ID":"c6747e4d-4e79-4199-942b-8354cbafa962","Type":"ContainerDied","Data":"a1b71887a857052978d270a767bb6122b203861b5ff1580a88cf508f98bcc19b"} Feb 18 15:09:46 crc kubenswrapper[4896]: I0218 15:09:46.892046 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-q2wgt" Feb 18 15:09:47 crc kubenswrapper[4896]: I0218 15:09:47.849329 4896 generic.go:334] "Generic (PLEG): container finished" podID="c6747e4d-4e79-4199-942b-8354cbafa962" containerID="0d4951fe949e403398c42bb14f45a7d181a253ecb063bbca01f74e694501efe1" exitCode=0 Feb 18 15:09:47 crc kubenswrapper[4896]: I0218 15:09:47.849433 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp" event={"ID":"c6747e4d-4e79-4199-942b-8354cbafa962","Type":"ContainerDied","Data":"0d4951fe949e403398c42bb14f45a7d181a253ecb063bbca01f74e694501efe1"} Feb 18 15:09:49 crc kubenswrapper[4896]: I0218 15:09:49.127767 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp" Feb 18 15:09:49 crc kubenswrapper[4896]: I0218 15:09:49.229482 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c6747e4d-4e79-4199-942b-8354cbafa962-util\") pod \"c6747e4d-4e79-4199-942b-8354cbafa962\" (UID: \"c6747e4d-4e79-4199-942b-8354cbafa962\") " Feb 18 15:09:49 crc kubenswrapper[4896]: I0218 15:09:49.229595 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c6747e4d-4e79-4199-942b-8354cbafa962-bundle\") pod \"c6747e4d-4e79-4199-942b-8354cbafa962\" (UID: \"c6747e4d-4e79-4199-942b-8354cbafa962\") " Feb 18 15:09:49 crc kubenswrapper[4896]: I0218 15:09:49.229639 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpjj4\" (UniqueName: \"kubernetes.io/projected/c6747e4d-4e79-4199-942b-8354cbafa962-kube-api-access-bpjj4\") pod \"c6747e4d-4e79-4199-942b-8354cbafa962\" (UID: \"c6747e4d-4e79-4199-942b-8354cbafa962\") " Feb 18 15:09:49 crc kubenswrapper[4896]: I0218 15:09:49.230203 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6747e4d-4e79-4199-942b-8354cbafa962-bundle" (OuterVolumeSpecName: "bundle") pod "c6747e4d-4e79-4199-942b-8354cbafa962" (UID: "c6747e4d-4e79-4199-942b-8354cbafa962"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:09:49 crc kubenswrapper[4896]: I0218 15:09:49.234043 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6747e4d-4e79-4199-942b-8354cbafa962-kube-api-access-bpjj4" (OuterVolumeSpecName: "kube-api-access-bpjj4") pod "c6747e4d-4e79-4199-942b-8354cbafa962" (UID: "c6747e4d-4e79-4199-942b-8354cbafa962"). InnerVolumeSpecName "kube-api-access-bpjj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:09:49 crc kubenswrapper[4896]: I0218 15:09:49.243680 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6747e4d-4e79-4199-942b-8354cbafa962-util" (OuterVolumeSpecName: "util") pod "c6747e4d-4e79-4199-942b-8354cbafa962" (UID: "c6747e4d-4e79-4199-942b-8354cbafa962"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:09:49 crc kubenswrapper[4896]: I0218 15:09:49.331411 4896 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c6747e4d-4e79-4199-942b-8354cbafa962-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 15:09:49 crc kubenswrapper[4896]: I0218 15:09:49.331442 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpjj4\" (UniqueName: \"kubernetes.io/projected/c6747e4d-4e79-4199-942b-8354cbafa962-kube-api-access-bpjj4\") on node \"crc\" DevicePath \"\"" Feb 18 15:09:49 crc kubenswrapper[4896]: I0218 15:09:49.331453 4896 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c6747e4d-4e79-4199-942b-8354cbafa962-util\") on node \"crc\" DevicePath \"\"" Feb 18 15:09:49 crc kubenswrapper[4896]: I0218 15:09:49.861356 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp" event={"ID":"c6747e4d-4e79-4199-942b-8354cbafa962","Type":"ContainerDied","Data":"09751550d2bce355911ac793f754482942f6dc6cddc451bff768d592dddb2b9b"} Feb 18 15:09:49 crc kubenswrapper[4896]: I0218 15:09:49.861399 4896 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="09751550d2bce355911ac793f754482942f6dc6cddc451bff768d592dddb2b9b" Feb 18 15:09:49 crc kubenswrapper[4896]: I0218 15:09:49.861440 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp" Feb 18 15:09:50 crc kubenswrapper[4896]: I0218 15:09:50.677699 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q2wgt"] Feb 18 15:09:50 crc kubenswrapper[4896]: I0218 15:09:50.677900 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-q2wgt" podUID="3b1335f8-9329-48fa-b924-f6729c81100d" containerName="registry-server" containerID="cri-o://f370806378c462f1b8f1109c1095a8e944e4dc74ad68fe67da21be95e62c0ab7" gracePeriod=2 Feb 18 15:09:50 crc kubenswrapper[4896]: I0218 15:09:50.869258 4896 generic.go:334] "Generic (PLEG): container finished" podID="3b1335f8-9329-48fa-b924-f6729c81100d" containerID="f370806378c462f1b8f1109c1095a8e944e4dc74ad68fe67da21be95e62c0ab7" exitCode=0 Feb 18 15:09:50 crc kubenswrapper[4896]: I0218 15:09:50.869309 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q2wgt" event={"ID":"3b1335f8-9329-48fa-b924-f6729c81100d","Type":"ContainerDied","Data":"f370806378c462f1b8f1109c1095a8e944e4dc74ad68fe67da21be95e62c0ab7"} Feb 18 15:09:51 crc kubenswrapper[4896]: I0218 15:09:51.107512 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q2wgt" Feb 18 15:09:51 crc kubenswrapper[4896]: I0218 15:09:51.154012 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b1335f8-9329-48fa-b924-f6729c81100d-catalog-content\") pod \"3b1335f8-9329-48fa-b924-f6729c81100d\" (UID: \"3b1335f8-9329-48fa-b924-f6729c81100d\") " Feb 18 15:09:51 crc kubenswrapper[4896]: I0218 15:09:51.154093 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x42dz\" (UniqueName: \"kubernetes.io/projected/3b1335f8-9329-48fa-b924-f6729c81100d-kube-api-access-x42dz\") pod \"3b1335f8-9329-48fa-b924-f6729c81100d\" (UID: \"3b1335f8-9329-48fa-b924-f6729c81100d\") " Feb 18 15:09:51 crc kubenswrapper[4896]: I0218 15:09:51.154146 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b1335f8-9329-48fa-b924-f6729c81100d-utilities\") pod \"3b1335f8-9329-48fa-b924-f6729c81100d\" (UID: \"3b1335f8-9329-48fa-b924-f6729c81100d\") " Feb 18 15:09:51 crc kubenswrapper[4896]: I0218 15:09:51.155116 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b1335f8-9329-48fa-b924-f6729c81100d-utilities" (OuterVolumeSpecName: "utilities") pod "3b1335f8-9329-48fa-b924-f6729c81100d" (UID: "3b1335f8-9329-48fa-b924-f6729c81100d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:09:51 crc kubenswrapper[4896]: I0218 15:09:51.164336 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b1335f8-9329-48fa-b924-f6729c81100d-kube-api-access-x42dz" (OuterVolumeSpecName: "kube-api-access-x42dz") pod "3b1335f8-9329-48fa-b924-f6729c81100d" (UID: "3b1335f8-9329-48fa-b924-f6729c81100d"). InnerVolumeSpecName "kube-api-access-x42dz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:09:51 crc kubenswrapper[4896]: I0218 15:09:51.199967 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b1335f8-9329-48fa-b924-f6729c81100d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3b1335f8-9329-48fa-b924-f6729c81100d" (UID: "3b1335f8-9329-48fa-b924-f6729c81100d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:09:51 crc kubenswrapper[4896]: I0218 15:09:51.255852 4896 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b1335f8-9329-48fa-b924-f6729c81100d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 15:09:51 crc kubenswrapper[4896]: I0218 15:09:51.255899 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x42dz\" (UniqueName: \"kubernetes.io/projected/3b1335f8-9329-48fa-b924-f6729c81100d-kube-api-access-x42dz\") on node \"crc\" DevicePath \"\"" Feb 18 15:09:51 crc kubenswrapper[4896]: I0218 15:09:51.255918 4896 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b1335f8-9329-48fa-b924-f6729c81100d-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 15:09:51 crc kubenswrapper[4896]: I0218 15:09:51.876010 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q2wgt" event={"ID":"3b1335f8-9329-48fa-b924-f6729c81100d","Type":"ContainerDied","Data":"551e7d2d0fc16f626994bbf110f94ffe8390851a7ad4ba85f5ac5e23f557421c"} Feb 18 15:09:51 crc kubenswrapper[4896]: I0218 15:09:51.877065 4896 scope.go:117] "RemoveContainer" containerID="f370806378c462f1b8f1109c1095a8e944e4dc74ad68fe67da21be95e62c0ab7" Feb 18 15:09:51 crc kubenswrapper[4896]: I0218 15:09:51.876068 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q2wgt" Feb 18 15:09:51 crc kubenswrapper[4896]: I0218 15:09:51.905774 4896 scope.go:117] "RemoveContainer" containerID="b8fcbaade3af77067e9703169f6034047980a492e605b2b8bc7213849859945b" Feb 18 15:09:51 crc kubenswrapper[4896]: I0218 15:09:51.913962 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q2wgt"] Feb 18 15:09:51 crc kubenswrapper[4896]: I0218 15:09:51.918701 4896 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-q2wgt"] Feb 18 15:09:51 crc kubenswrapper[4896]: I0218 15:09:51.923428 4896 scope.go:117] "RemoveContainer" containerID="674325e530bd8a879c64f8a1728ee47d9044472076db2451ff2017c549f26f1a" Feb 18 15:09:51 crc kubenswrapper[4896]: I0218 15:09:51.937471 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b1335f8-9329-48fa-b924-f6729c81100d" path="/var/lib/kubelet/pods/3b1335f8-9329-48fa-b924-f6729c81100d/volumes" Feb 18 15:09:56 crc kubenswrapper[4896]: I0218 15:09:56.177612 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-v9qdn"] Feb 18 15:09:56 crc kubenswrapper[4896]: E0218 15:09:56.178153 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b1335f8-9329-48fa-b924-f6729c81100d" containerName="registry-server" Feb 18 15:09:56 crc kubenswrapper[4896]: I0218 15:09:56.178168 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b1335f8-9329-48fa-b924-f6729c81100d" containerName="registry-server" Feb 18 15:09:56 crc kubenswrapper[4896]: E0218 15:09:56.178185 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6747e4d-4e79-4199-942b-8354cbafa962" containerName="pull" Feb 18 15:09:56 crc kubenswrapper[4896]: I0218 15:09:56.178192 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6747e4d-4e79-4199-942b-8354cbafa962" containerName="pull" Feb 18 15:09:56 crc kubenswrapper[4896]: E0218 15:09:56.178226 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b1335f8-9329-48fa-b924-f6729c81100d" containerName="extract-content" Feb 18 15:09:56 crc kubenswrapper[4896]: I0218 15:09:56.178252 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b1335f8-9329-48fa-b924-f6729c81100d" containerName="extract-content" Feb 18 15:09:56 crc kubenswrapper[4896]: E0218 15:09:56.178262 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b1335f8-9329-48fa-b924-f6729c81100d" containerName="extract-utilities" Feb 18 15:09:56 crc kubenswrapper[4896]: I0218 15:09:56.178269 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b1335f8-9329-48fa-b924-f6729c81100d" containerName="extract-utilities" Feb 18 15:09:56 crc kubenswrapper[4896]: E0218 15:09:56.178280 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6747e4d-4e79-4199-942b-8354cbafa962" containerName="extract" Feb 18 15:09:56 crc kubenswrapper[4896]: I0218 15:09:56.178288 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6747e4d-4e79-4199-942b-8354cbafa962" containerName="extract" Feb 18 15:09:56 crc kubenswrapper[4896]: E0218 15:09:56.178299 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6747e4d-4e79-4199-942b-8354cbafa962" containerName="util" Feb 18 15:09:56 crc kubenswrapper[4896]: I0218 15:09:56.178306 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6747e4d-4e79-4199-942b-8354cbafa962" containerName="util" Feb 18 15:09:56 crc kubenswrapper[4896]: I0218 15:09:56.178434 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6747e4d-4e79-4199-942b-8354cbafa962" containerName="extract" Feb 18 15:09:56 crc kubenswrapper[4896]: I0218 15:09:56.178445 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b1335f8-9329-48fa-b924-f6729c81100d" containerName="registry-server" Feb 18 15:09:56 crc kubenswrapper[4896]: I0218 15:09:56.178816 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-v9qdn" Feb 18 15:09:56 crc kubenswrapper[4896]: I0218 15:09:56.180755 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-dockercfg-8hd4t" Feb 18 15:09:56 crc kubenswrapper[4896]: I0218 15:09:56.192143 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-v9qdn"] Feb 18 15:09:56 crc kubenswrapper[4896]: I0218 15:09:56.311924 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7r9fj\" (UniqueName: \"kubernetes.io/projected/ddd714d9-66f9-4caa-8d0c-a6e0ff935af0-kube-api-access-7r9fj\") pod \"rabbitmq-cluster-operator-779fc9694b-v9qdn\" (UID: \"ddd714d9-66f9-4caa-8d0c-a6e0ff935af0\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-v9qdn" Feb 18 15:09:56 crc kubenswrapper[4896]: I0218 15:09:56.413059 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7r9fj\" (UniqueName: \"kubernetes.io/projected/ddd714d9-66f9-4caa-8d0c-a6e0ff935af0-kube-api-access-7r9fj\") pod \"rabbitmq-cluster-operator-779fc9694b-v9qdn\" (UID: \"ddd714d9-66f9-4caa-8d0c-a6e0ff935af0\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-v9qdn" Feb 18 15:09:56 crc kubenswrapper[4896]: I0218 15:09:56.434332 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7r9fj\" (UniqueName: \"kubernetes.io/projected/ddd714d9-66f9-4caa-8d0c-a6e0ff935af0-kube-api-access-7r9fj\") pod \"rabbitmq-cluster-operator-779fc9694b-v9qdn\" (UID: \"ddd714d9-66f9-4caa-8d0c-a6e0ff935af0\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-v9qdn" Feb 18 15:09:56 crc kubenswrapper[4896]: I0218 15:09:56.495652 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-v9qdn" Feb 18 15:09:56 crc kubenswrapper[4896]: I0218 15:09:56.737528 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-v9qdn"] Feb 18 15:09:56 crc kubenswrapper[4896]: I0218 15:09:56.903713 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-v9qdn" event={"ID":"ddd714d9-66f9-4caa-8d0c-a6e0ff935af0","Type":"ContainerStarted","Data":"bf49fe940504408e1b1a3e483766e18c24266944beabce6cf17789c8b89701f6"} Feb 18 15:10:00 crc kubenswrapper[4896]: I0218 15:10:00.478397 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-774gb"] Feb 18 15:10:00 crc kubenswrapper[4896]: I0218 15:10:00.480174 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-774gb" Feb 18 15:10:00 crc kubenswrapper[4896]: I0218 15:10:00.500109 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-774gb"] Feb 18 15:10:00 crc kubenswrapper[4896]: I0218 15:10:00.570679 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5-utilities\") pod \"community-operators-774gb\" (UID: \"2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5\") " pod="openshift-marketplace/community-operators-774gb" Feb 18 15:10:00 crc kubenswrapper[4896]: I0218 15:10:00.570736 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gm8dc\" (UniqueName: \"kubernetes.io/projected/2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5-kube-api-access-gm8dc\") pod \"community-operators-774gb\" (UID: \"2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5\") " pod="openshift-marketplace/community-operators-774gb" Feb 18 15:10:00 crc kubenswrapper[4896]: I0218 15:10:00.570775 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5-catalog-content\") pod \"community-operators-774gb\" (UID: \"2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5\") " pod="openshift-marketplace/community-operators-774gb" Feb 18 15:10:00 crc kubenswrapper[4896]: I0218 15:10:00.672315 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5-utilities\") pod \"community-operators-774gb\" (UID: \"2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5\") " pod="openshift-marketplace/community-operators-774gb" Feb 18 15:10:00 crc kubenswrapper[4896]: I0218 15:10:00.672369 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gm8dc\" (UniqueName: \"kubernetes.io/projected/2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5-kube-api-access-gm8dc\") pod \"community-operators-774gb\" (UID: \"2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5\") " pod="openshift-marketplace/community-operators-774gb" Feb 18 15:10:00 crc kubenswrapper[4896]: I0218 15:10:00.672398 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5-catalog-content\") pod \"community-operators-774gb\" (UID: \"2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5\") " pod="openshift-marketplace/community-operators-774gb" Feb 18 15:10:00 crc kubenswrapper[4896]: I0218 15:10:00.672803 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5-catalog-content\") pod \"community-operators-774gb\" (UID: \"2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5\") " pod="openshift-marketplace/community-operators-774gb" Feb 18 15:10:00 crc kubenswrapper[4896]: I0218 15:10:00.672882 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5-utilities\") pod \"community-operators-774gb\" (UID: \"2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5\") " pod="openshift-marketplace/community-operators-774gb" Feb 18 15:10:00 crc kubenswrapper[4896]: I0218 15:10:00.693683 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gm8dc\" (UniqueName: \"kubernetes.io/projected/2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5-kube-api-access-gm8dc\") pod \"community-operators-774gb\" (UID: \"2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5\") " pod="openshift-marketplace/community-operators-774gb" Feb 18 15:10:00 crc kubenswrapper[4896]: I0218 15:10:00.793777 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-774gb" Feb 18 15:10:00 crc kubenswrapper[4896]: I0218 15:10:00.940500 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-v9qdn" event={"ID":"ddd714d9-66f9-4caa-8d0c-a6e0ff935af0","Type":"ContainerStarted","Data":"94729f4366d5a249c60f6e0422b880a0f5053af00c99916f052b66983d08f27e"} Feb 18 15:10:01 crc kubenswrapper[4896]: I0218 15:10:01.163277 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-v9qdn" podStartSLOduration=1.7323365069999999 podStartE2EDuration="5.163262614s" podCreationTimestamp="2026-02-18 15:09:56 +0000 UTC" firstStartedPulling="2026-02-18 15:09:56.742595672 +0000 UTC m=+895.671929885" lastFinishedPulling="2026-02-18 15:10:00.173521779 +0000 UTC m=+899.102855992" observedRunningTime="2026-02-18 15:10:00.98340222 +0000 UTC m=+899.912736443" watchObservedRunningTime="2026-02-18 15:10:01.163262614 +0000 UTC m=+900.092596827" Feb 18 15:10:01 crc kubenswrapper[4896]: I0218 15:10:01.166374 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-774gb"] Feb 18 15:10:01 crc kubenswrapper[4896]: W0218 15:10:01.174796 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2cf7f60d_2aae_49e4_84af_1fd5c98b0ea5.slice/crio-5fe91d8f6605b94a801d0ca4a671fdadfe890713f6cc43c215f8dd81473bc430 WatchSource:0}: Error finding container 5fe91d8f6605b94a801d0ca4a671fdadfe890713f6cc43c215f8dd81473bc430: Status 404 returned error can't find the container with id 5fe91d8f6605b94a801d0ca4a671fdadfe890713f6cc43c215f8dd81473bc430 Feb 18 15:10:01 crc kubenswrapper[4896]: I0218 15:10:01.951970 4896 generic.go:334] "Generic (PLEG): container finished" podID="2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5" containerID="28a78da58db6a551ea3e092bd75c48b316e249a50ba3a9cdea85901813015c9a" exitCode=0 Feb 18 15:10:01 crc kubenswrapper[4896]: I0218 15:10:01.952048 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-774gb" event={"ID":"2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5","Type":"ContainerDied","Data":"28a78da58db6a551ea3e092bd75c48b316e249a50ba3a9cdea85901813015c9a"} Feb 18 15:10:01 crc kubenswrapper[4896]: I0218 15:10:01.952384 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-774gb" event={"ID":"2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5","Type":"ContainerStarted","Data":"5fe91d8f6605b94a801d0ca4a671fdadfe890713f6cc43c215f8dd81473bc430"} Feb 18 15:10:03 crc kubenswrapper[4896]: I0218 15:10:03.963818 4896 generic.go:334] "Generic (PLEG): container finished" podID="2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5" containerID="19ff8cd716b54e263661cc720c653da29482ee03861249965547a348b58a9065" exitCode=0 Feb 18 15:10:03 crc kubenswrapper[4896]: I0218 15:10:03.963907 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-774gb" event={"ID":"2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5","Type":"ContainerDied","Data":"19ff8cd716b54e263661cc720c653da29482ee03861249965547a348b58a9065"} Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.050850 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["test-operator-kuttl-tests/rabbitmq-server-0"] Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.052062 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="test-operator-kuttl-tests/rabbitmq-server-0" Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.057839 4896 reflector.go:368] Caches populated for *v1.Secret from object-"test-operator-kuttl-tests"/"rabbitmq-server-dockercfg-ncs2l" Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.057875 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"test-operator-kuttl-tests"/"rabbitmq-plugins-conf" Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.057845 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"test-operator-kuttl-tests"/"rabbitmq-server-conf" Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.058023 4896 reflector.go:368] Caches populated for *v1.Secret from object-"test-operator-kuttl-tests"/"rabbitmq-erlang-cookie" Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.058099 4896 reflector.go:368] Caches populated for *v1.Secret from object-"test-operator-kuttl-tests"/"rabbitmq-default-user" Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.058184 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["test-operator-kuttl-tests/rabbitmq-server-0"] Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.076828 4896 patch_prober.go:28] interesting pod/machine-config-daemon-24rq5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.077121 4896 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.221223 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c5ce908c-bcd7-456e-9ab2-abe730180641-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c5ce908c-bcd7-456e-9ab2-abe730180641\") " pod="test-operator-kuttl-tests/rabbitmq-server-0" Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.221336 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c5ce908c-bcd7-456e-9ab2-abe730180641-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c5ce908c-bcd7-456e-9ab2-abe730180641\") " pod="test-operator-kuttl-tests/rabbitmq-server-0" Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.221361 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c5ce908c-bcd7-456e-9ab2-abe730180641-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c5ce908c-bcd7-456e-9ab2-abe730180641\") " pod="test-operator-kuttl-tests/rabbitmq-server-0" Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.221449 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-beb69ef9-f4b9-4aca-ad02-9073a6b39723\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-beb69ef9-f4b9-4aca-ad02-9073a6b39723\") pod \"rabbitmq-server-0\" (UID: \"c5ce908c-bcd7-456e-9ab2-abe730180641\") " pod="test-operator-kuttl-tests/rabbitmq-server-0" Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.221468 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c5ce908c-bcd7-456e-9ab2-abe730180641-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c5ce908c-bcd7-456e-9ab2-abe730180641\") " pod="test-operator-kuttl-tests/rabbitmq-server-0" Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.221491 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxwmm\" (UniqueName: \"kubernetes.io/projected/c5ce908c-bcd7-456e-9ab2-abe730180641-kube-api-access-wxwmm\") pod \"rabbitmq-server-0\" (UID: \"c5ce908c-bcd7-456e-9ab2-abe730180641\") " pod="test-operator-kuttl-tests/rabbitmq-server-0" Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.221512 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c5ce908c-bcd7-456e-9ab2-abe730180641-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c5ce908c-bcd7-456e-9ab2-abe730180641\") " pod="test-operator-kuttl-tests/rabbitmq-server-0" Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.221531 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c5ce908c-bcd7-456e-9ab2-abe730180641-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c5ce908c-bcd7-456e-9ab2-abe730180641\") " pod="test-operator-kuttl-tests/rabbitmq-server-0" Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.322630 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c5ce908c-bcd7-456e-9ab2-abe730180641-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c5ce908c-bcd7-456e-9ab2-abe730180641\") " pod="test-operator-kuttl-tests/rabbitmq-server-0" Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.322679 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c5ce908c-bcd7-456e-9ab2-abe730180641-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c5ce908c-bcd7-456e-9ab2-abe730180641\") " pod="test-operator-kuttl-tests/rabbitmq-server-0" Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.322701 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c5ce908c-bcd7-456e-9ab2-abe730180641-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c5ce908c-bcd7-456e-9ab2-abe730180641\") " pod="test-operator-kuttl-tests/rabbitmq-server-0" Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.322729 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c5ce908c-bcd7-456e-9ab2-abe730180641-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c5ce908c-bcd7-456e-9ab2-abe730180641\") " pod="test-operator-kuttl-tests/rabbitmq-server-0" Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.322749 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-beb69ef9-f4b9-4aca-ad02-9073a6b39723\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-beb69ef9-f4b9-4aca-ad02-9073a6b39723\") pod \"rabbitmq-server-0\" (UID: \"c5ce908c-bcd7-456e-9ab2-abe730180641\") " pod="test-operator-kuttl-tests/rabbitmq-server-0" Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.322769 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxwmm\" (UniqueName: \"kubernetes.io/projected/c5ce908c-bcd7-456e-9ab2-abe730180641-kube-api-access-wxwmm\") pod \"rabbitmq-server-0\" (UID: \"c5ce908c-bcd7-456e-9ab2-abe730180641\") " pod="test-operator-kuttl-tests/rabbitmq-server-0" Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.322795 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c5ce908c-bcd7-456e-9ab2-abe730180641-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c5ce908c-bcd7-456e-9ab2-abe730180641\") " pod="test-operator-kuttl-tests/rabbitmq-server-0" Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.322815 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c5ce908c-bcd7-456e-9ab2-abe730180641-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c5ce908c-bcd7-456e-9ab2-abe730180641\") " pod="test-operator-kuttl-tests/rabbitmq-server-0" Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.323519 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c5ce908c-bcd7-456e-9ab2-abe730180641-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c5ce908c-bcd7-456e-9ab2-abe730180641\") " pod="test-operator-kuttl-tests/rabbitmq-server-0" Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.323904 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c5ce908c-bcd7-456e-9ab2-abe730180641-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c5ce908c-bcd7-456e-9ab2-abe730180641\") " pod="test-operator-kuttl-tests/rabbitmq-server-0" Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.324080 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c5ce908c-bcd7-456e-9ab2-abe730180641-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c5ce908c-bcd7-456e-9ab2-abe730180641\") " pod="test-operator-kuttl-tests/rabbitmq-server-0" Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.327821 4896 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.327863 4896 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-beb69ef9-f4b9-4aca-ad02-9073a6b39723\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-beb69ef9-f4b9-4aca-ad02-9073a6b39723\") pod \"rabbitmq-server-0\" (UID: \"c5ce908c-bcd7-456e-9ab2-abe730180641\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/90653de0ec125d495bc04579f76d2ccd7e7aae7633c0992922ae99853fbf6fe0/globalmount\"" pod="test-operator-kuttl-tests/rabbitmq-server-0" Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.328087 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c5ce908c-bcd7-456e-9ab2-abe730180641-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c5ce908c-bcd7-456e-9ab2-abe730180641\") " pod="test-operator-kuttl-tests/rabbitmq-server-0" Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.330081 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c5ce908c-bcd7-456e-9ab2-abe730180641-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c5ce908c-bcd7-456e-9ab2-abe730180641\") " pod="test-operator-kuttl-tests/rabbitmq-server-0" Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.336154 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c5ce908c-bcd7-456e-9ab2-abe730180641-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c5ce908c-bcd7-456e-9ab2-abe730180641\") " pod="test-operator-kuttl-tests/rabbitmq-server-0" Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.346351 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxwmm\" (UniqueName: \"kubernetes.io/projected/c5ce908c-bcd7-456e-9ab2-abe730180641-kube-api-access-wxwmm\") pod \"rabbitmq-server-0\" (UID: \"c5ce908c-bcd7-456e-9ab2-abe730180641\") " pod="test-operator-kuttl-tests/rabbitmq-server-0" Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.349352 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-beb69ef9-f4b9-4aca-ad02-9073a6b39723\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-beb69ef9-f4b9-4aca-ad02-9073a6b39723\") pod \"rabbitmq-server-0\" (UID: \"c5ce908c-bcd7-456e-9ab2-abe730180641\") " pod="test-operator-kuttl-tests/rabbitmq-server-0" Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.381122 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="test-operator-kuttl-tests/rabbitmq-server-0" Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.808412 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["test-operator-kuttl-tests/rabbitmq-server-0"] Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.987298 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-774gb" event={"ID":"2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5","Type":"ContainerStarted","Data":"4ad0152236ab99fa7a3fac595af67c432c3795c78a5be7102d4eb1d61ede7e32"} Feb 18 15:10:04 crc kubenswrapper[4896]: I0218 15:10:04.989506 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/rabbitmq-server-0" event={"ID":"c5ce908c-bcd7-456e-9ab2-abe730180641","Type":"ContainerStarted","Data":"98e72d8e0343b24dba309968057d3de22ec325818b3e23873cfac18a3ff8ef76"} Feb 18 15:10:05 crc kubenswrapper[4896]: I0218 15:10:05.007118 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-774gb" podStartSLOduration=2.594292698 podStartE2EDuration="5.007098699s" podCreationTimestamp="2026-02-18 15:10:00 +0000 UTC" firstStartedPulling="2026-02-18 15:10:01.956255029 +0000 UTC m=+900.885589242" lastFinishedPulling="2026-02-18 15:10:04.36906101 +0000 UTC m=+903.298395243" observedRunningTime="2026-02-18 15:10:05.003173515 +0000 UTC m=+903.932507728" watchObservedRunningTime="2026-02-18 15:10:05.007098699 +0000 UTC m=+903.936432932" Feb 18 15:10:06 crc kubenswrapper[4896]: I0218 15:10:06.676149 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-bgmmh"] Feb 18 15:10:06 crc kubenswrapper[4896]: I0218 15:10:06.677445 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-bgmmh" Feb 18 15:10:06 crc kubenswrapper[4896]: I0218 15:10:06.682094 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-index-dockercfg-zt5pc" Feb 18 15:10:06 crc kubenswrapper[4896]: I0218 15:10:06.690948 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-bgmmh"] Feb 18 15:10:06 crc kubenswrapper[4896]: I0218 15:10:06.752742 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qt4z\" (UniqueName: \"kubernetes.io/projected/75da0d85-5bc2-41b4-8502-edd5b7dfd6ea-kube-api-access-4qt4z\") pod \"keystone-operator-index-bgmmh\" (UID: \"75da0d85-5bc2-41b4-8502-edd5b7dfd6ea\") " pod="openstack-operators/keystone-operator-index-bgmmh" Feb 18 15:10:06 crc kubenswrapper[4896]: I0218 15:10:06.853939 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qt4z\" (UniqueName: \"kubernetes.io/projected/75da0d85-5bc2-41b4-8502-edd5b7dfd6ea-kube-api-access-4qt4z\") pod \"keystone-operator-index-bgmmh\" (UID: \"75da0d85-5bc2-41b4-8502-edd5b7dfd6ea\") " pod="openstack-operators/keystone-operator-index-bgmmh" Feb 18 15:10:06 crc kubenswrapper[4896]: I0218 15:10:06.872226 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qt4z\" (UniqueName: \"kubernetes.io/projected/75da0d85-5bc2-41b4-8502-edd5b7dfd6ea-kube-api-access-4qt4z\") pod \"keystone-operator-index-bgmmh\" (UID: \"75da0d85-5bc2-41b4-8502-edd5b7dfd6ea\") " pod="openstack-operators/keystone-operator-index-bgmmh" Feb 18 15:10:07 crc kubenswrapper[4896]: I0218 15:10:07.006424 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-bgmmh" Feb 18 15:10:09 crc kubenswrapper[4896]: I0218 15:10:09.042541 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-bgmmh"] Feb 18 15:10:09 crc kubenswrapper[4896]: W0218 15:10:09.051785 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75da0d85_5bc2_41b4_8502_edd5b7dfd6ea.slice/crio-3b703ea44b521f9dd66156ce3fa3fe6bcbddb43f21dc08fdb9ecac1acb3a488f WatchSource:0}: Error finding container 3b703ea44b521f9dd66156ce3fa3fe6bcbddb43f21dc08fdb9ecac1acb3a488f: Status 404 returned error can't find the container with id 3b703ea44b521f9dd66156ce3fa3fe6bcbddb43f21dc08fdb9ecac1acb3a488f Feb 18 15:10:09 crc kubenswrapper[4896]: I0218 15:10:09.184440 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-bgmmh" event={"ID":"75da0d85-5bc2-41b4-8502-edd5b7dfd6ea","Type":"ContainerStarted","Data":"3b703ea44b521f9dd66156ce3fa3fe6bcbddb43f21dc08fdb9ecac1acb3a488f"} Feb 18 15:10:10 crc kubenswrapper[4896]: I0218 15:10:10.794342 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-774gb" Feb 18 15:10:10 crc kubenswrapper[4896]: I0218 15:10:10.795577 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-774gb" Feb 18 15:10:10 crc kubenswrapper[4896]: I0218 15:10:10.850069 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-774gb" Feb 18 15:10:11 crc kubenswrapper[4896]: I0218 15:10:11.251555 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-774gb" Feb 18 15:10:12 crc kubenswrapper[4896]: I0218 15:10:12.070949 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-774gb"] Feb 18 15:10:13 crc kubenswrapper[4896]: I0218 15:10:13.218143 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-774gb" podUID="2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5" containerName="registry-server" containerID="cri-o://4ad0152236ab99fa7a3fac595af67c432c3795c78a5be7102d4eb1d61ede7e32" gracePeriod=2 Feb 18 15:10:13 crc kubenswrapper[4896]: I0218 15:10:13.675368 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-774gb" Feb 18 15:10:13 crc kubenswrapper[4896]: I0218 15:10:13.750036 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gm8dc\" (UniqueName: \"kubernetes.io/projected/2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5-kube-api-access-gm8dc\") pod \"2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5\" (UID: \"2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5\") " Feb 18 15:10:13 crc kubenswrapper[4896]: I0218 15:10:13.750452 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5-catalog-content\") pod \"2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5\" (UID: \"2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5\") " Feb 18 15:10:13 crc kubenswrapper[4896]: I0218 15:10:13.750600 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5-utilities\") pod \"2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5\" (UID: \"2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5\") " Feb 18 15:10:13 crc kubenswrapper[4896]: I0218 15:10:13.751368 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5-utilities" (OuterVolumeSpecName: "utilities") pod "2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5" (UID: "2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:10:13 crc kubenswrapper[4896]: I0218 15:10:13.848884 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5-kube-api-access-gm8dc" (OuterVolumeSpecName: "kube-api-access-gm8dc") pod "2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5" (UID: "2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5"). InnerVolumeSpecName "kube-api-access-gm8dc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:10:13 crc kubenswrapper[4896]: I0218 15:10:13.851793 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gm8dc\" (UniqueName: \"kubernetes.io/projected/2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5-kube-api-access-gm8dc\") on node \"crc\" DevicePath \"\"" Feb 18 15:10:13 crc kubenswrapper[4896]: I0218 15:10:13.851830 4896 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 15:10:14 crc kubenswrapper[4896]: I0218 15:10:14.224303 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-bgmmh" event={"ID":"75da0d85-5bc2-41b4-8502-edd5b7dfd6ea","Type":"ContainerStarted","Data":"ec4fe34586193c59587feb12e6ce3d53d2936c8c7b1c402046baacd81ad9083b"} Feb 18 15:10:14 crc kubenswrapper[4896]: I0218 15:10:14.227358 4896 generic.go:334] "Generic (PLEG): container finished" podID="2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5" containerID="4ad0152236ab99fa7a3fac595af67c432c3795c78a5be7102d4eb1d61ede7e32" exitCode=0 Feb 18 15:10:14 crc kubenswrapper[4896]: I0218 15:10:14.227394 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-774gb" event={"ID":"2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5","Type":"ContainerDied","Data":"4ad0152236ab99fa7a3fac595af67c432c3795c78a5be7102d4eb1d61ede7e32"} Feb 18 15:10:14 crc kubenswrapper[4896]: I0218 15:10:14.227418 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-774gb" event={"ID":"2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5","Type":"ContainerDied","Data":"5fe91d8f6605b94a801d0ca4a671fdadfe890713f6cc43c215f8dd81473bc430"} Feb 18 15:10:14 crc kubenswrapper[4896]: I0218 15:10:14.227439 4896 scope.go:117] "RemoveContainer" containerID="4ad0152236ab99fa7a3fac595af67c432c3795c78a5be7102d4eb1d61ede7e32" Feb 18 15:10:14 crc kubenswrapper[4896]: I0218 15:10:14.227554 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-774gb" Feb 18 15:10:14 crc kubenswrapper[4896]: I0218 15:10:14.243215 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-bgmmh" podStartSLOduration=3.834440175 podStartE2EDuration="8.243199413s" podCreationTimestamp="2026-02-18 15:10:06 +0000 UTC" firstStartedPulling="2026-02-18 15:10:09.05314212 +0000 UTC m=+907.982476333" lastFinishedPulling="2026-02-18 15:10:13.461901358 +0000 UTC m=+912.391235571" observedRunningTime="2026-02-18 15:10:14.239976648 +0000 UTC m=+913.169310881" watchObservedRunningTime="2026-02-18 15:10:14.243199413 +0000 UTC m=+913.172533626" Feb 18 15:10:14 crc kubenswrapper[4896]: I0218 15:10:14.245494 4896 scope.go:117] "RemoveContainer" containerID="19ff8cd716b54e263661cc720c653da29482ee03861249965547a348b58a9065" Feb 18 15:10:14 crc kubenswrapper[4896]: I0218 15:10:14.259904 4896 scope.go:117] "RemoveContainer" containerID="28a78da58db6a551ea3e092bd75c48b316e249a50ba3a9cdea85901813015c9a" Feb 18 15:10:14 crc kubenswrapper[4896]: I0218 15:10:14.281081 4896 scope.go:117] "RemoveContainer" containerID="4ad0152236ab99fa7a3fac595af67c432c3795c78a5be7102d4eb1d61ede7e32" Feb 18 15:10:14 crc kubenswrapper[4896]: E0218 15:10:14.281554 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ad0152236ab99fa7a3fac595af67c432c3795c78a5be7102d4eb1d61ede7e32\": container with ID starting with 4ad0152236ab99fa7a3fac595af67c432c3795c78a5be7102d4eb1d61ede7e32 not found: ID does not exist" containerID="4ad0152236ab99fa7a3fac595af67c432c3795c78a5be7102d4eb1d61ede7e32" Feb 18 15:10:14 crc kubenswrapper[4896]: I0218 15:10:14.281602 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ad0152236ab99fa7a3fac595af67c432c3795c78a5be7102d4eb1d61ede7e32"} err="failed to get container status \"4ad0152236ab99fa7a3fac595af67c432c3795c78a5be7102d4eb1d61ede7e32\": rpc error: code = NotFound desc = could not find container \"4ad0152236ab99fa7a3fac595af67c432c3795c78a5be7102d4eb1d61ede7e32\": container with ID starting with 4ad0152236ab99fa7a3fac595af67c432c3795c78a5be7102d4eb1d61ede7e32 not found: ID does not exist" Feb 18 15:10:14 crc kubenswrapper[4896]: I0218 15:10:14.281634 4896 scope.go:117] "RemoveContainer" containerID="19ff8cd716b54e263661cc720c653da29482ee03861249965547a348b58a9065" Feb 18 15:10:14 crc kubenswrapper[4896]: E0218 15:10:14.281898 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19ff8cd716b54e263661cc720c653da29482ee03861249965547a348b58a9065\": container with ID starting with 19ff8cd716b54e263661cc720c653da29482ee03861249965547a348b58a9065 not found: ID does not exist" containerID="19ff8cd716b54e263661cc720c653da29482ee03861249965547a348b58a9065" Feb 18 15:10:14 crc kubenswrapper[4896]: I0218 15:10:14.281924 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19ff8cd716b54e263661cc720c653da29482ee03861249965547a348b58a9065"} err="failed to get container status \"19ff8cd716b54e263661cc720c653da29482ee03861249965547a348b58a9065\": rpc error: code = NotFound desc = could not find container \"19ff8cd716b54e263661cc720c653da29482ee03861249965547a348b58a9065\": container with ID starting with 19ff8cd716b54e263661cc720c653da29482ee03861249965547a348b58a9065 not found: ID does not exist" Feb 18 15:10:14 crc kubenswrapper[4896]: I0218 15:10:14.281942 4896 scope.go:117] "RemoveContainer" containerID="28a78da58db6a551ea3e092bd75c48b316e249a50ba3a9cdea85901813015c9a" Feb 18 15:10:14 crc kubenswrapper[4896]: E0218 15:10:14.282247 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28a78da58db6a551ea3e092bd75c48b316e249a50ba3a9cdea85901813015c9a\": container with ID starting with 28a78da58db6a551ea3e092bd75c48b316e249a50ba3a9cdea85901813015c9a not found: ID does not exist" containerID="28a78da58db6a551ea3e092bd75c48b316e249a50ba3a9cdea85901813015c9a" Feb 18 15:10:14 crc kubenswrapper[4896]: I0218 15:10:14.282283 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28a78da58db6a551ea3e092bd75c48b316e249a50ba3a9cdea85901813015c9a"} err="failed to get container status \"28a78da58db6a551ea3e092bd75c48b316e249a50ba3a9cdea85901813015c9a\": rpc error: code = NotFound desc = could not find container \"28a78da58db6a551ea3e092bd75c48b316e249a50ba3a9cdea85901813015c9a\": container with ID starting with 28a78da58db6a551ea3e092bd75c48b316e249a50ba3a9cdea85901813015c9a not found: ID does not exist" Feb 18 15:10:14 crc kubenswrapper[4896]: I0218 15:10:14.515464 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5" (UID: "2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:10:14 crc kubenswrapper[4896]: I0218 15:10:14.555544 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-774gb"] Feb 18 15:10:14 crc kubenswrapper[4896]: I0218 15:10:14.560644 4896 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-774gb"] Feb 18 15:10:14 crc kubenswrapper[4896]: I0218 15:10:14.562384 4896 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 15:10:15 crc kubenswrapper[4896]: I0218 15:10:15.234466 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/rabbitmq-server-0" event={"ID":"c5ce908c-bcd7-456e-9ab2-abe730180641","Type":"ContainerStarted","Data":"964908cec1ea5b2b53e165375d511f323d804bf9fc89abc3a0d6d01d9b60bbb3"} Feb 18 15:10:15 crc kubenswrapper[4896]: I0218 15:10:15.935831 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5" path="/var/lib/kubelet/pods/2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5/volumes" Feb 18 15:10:17 crc kubenswrapper[4896]: I0218 15:10:17.007678 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-index-bgmmh" Feb 18 15:10:17 crc kubenswrapper[4896]: I0218 15:10:17.008013 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/keystone-operator-index-bgmmh" Feb 18 15:10:17 crc kubenswrapper[4896]: I0218 15:10:17.038403 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/keystone-operator-index-bgmmh" Feb 18 15:10:27 crc kubenswrapper[4896]: I0218 15:10:27.040819 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-index-bgmmh" Feb 18 15:10:34 crc kubenswrapper[4896]: I0218 15:10:34.075941 4896 patch_prober.go:28] interesting pod/machine-config-daemon-24rq5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:10:34 crc kubenswrapper[4896]: I0218 15:10:34.076527 4896 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:10:35 crc kubenswrapper[4896]: I0218 15:10:35.685273 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dflvt"] Feb 18 15:10:35 crc kubenswrapper[4896]: E0218 15:10:35.685782 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5" containerName="registry-server" Feb 18 15:10:35 crc kubenswrapper[4896]: I0218 15:10:35.685793 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5" containerName="registry-server" Feb 18 15:10:35 crc kubenswrapper[4896]: E0218 15:10:35.685810 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5" containerName="extract-content" Feb 18 15:10:35 crc kubenswrapper[4896]: I0218 15:10:35.685816 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5" containerName="extract-content" Feb 18 15:10:35 crc kubenswrapper[4896]: E0218 15:10:35.685827 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5" containerName="extract-utilities" Feb 18 15:10:35 crc kubenswrapper[4896]: I0218 15:10:35.685833 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5" containerName="extract-utilities" Feb 18 15:10:35 crc kubenswrapper[4896]: I0218 15:10:35.685943 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cf7f60d-2aae-49e4-84af-1fd5c98b0ea5" containerName="registry-server" Feb 18 15:10:35 crc kubenswrapper[4896]: I0218 15:10:35.686801 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dflvt" Feb 18 15:10:35 crc kubenswrapper[4896]: I0218 15:10:35.693501 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dflvt"] Feb 18 15:10:35 crc kubenswrapper[4896]: I0218 15:10:35.846815 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f328bf3-bacb-4058-bb20-bc99dabaff31-catalog-content\") pod \"redhat-operators-dflvt\" (UID: \"9f328bf3-bacb-4058-bb20-bc99dabaff31\") " pod="openshift-marketplace/redhat-operators-dflvt" Feb 18 15:10:35 crc kubenswrapper[4896]: I0218 15:10:35.846897 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4mm2\" (UniqueName: \"kubernetes.io/projected/9f328bf3-bacb-4058-bb20-bc99dabaff31-kube-api-access-h4mm2\") pod \"redhat-operators-dflvt\" (UID: \"9f328bf3-bacb-4058-bb20-bc99dabaff31\") " pod="openshift-marketplace/redhat-operators-dflvt" Feb 18 15:10:35 crc kubenswrapper[4896]: I0218 15:10:35.846971 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f328bf3-bacb-4058-bb20-bc99dabaff31-utilities\") pod \"redhat-operators-dflvt\" (UID: \"9f328bf3-bacb-4058-bb20-bc99dabaff31\") " pod="openshift-marketplace/redhat-operators-dflvt" Feb 18 15:10:35 crc kubenswrapper[4896]: I0218 15:10:35.948346 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f328bf3-bacb-4058-bb20-bc99dabaff31-utilities\") pod \"redhat-operators-dflvt\" (UID: \"9f328bf3-bacb-4058-bb20-bc99dabaff31\") " pod="openshift-marketplace/redhat-operators-dflvt" Feb 18 15:10:35 crc kubenswrapper[4896]: I0218 15:10:35.948426 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f328bf3-bacb-4058-bb20-bc99dabaff31-catalog-content\") pod \"redhat-operators-dflvt\" (UID: \"9f328bf3-bacb-4058-bb20-bc99dabaff31\") " pod="openshift-marketplace/redhat-operators-dflvt" Feb 18 15:10:35 crc kubenswrapper[4896]: I0218 15:10:35.948466 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4mm2\" (UniqueName: \"kubernetes.io/projected/9f328bf3-bacb-4058-bb20-bc99dabaff31-kube-api-access-h4mm2\") pod \"redhat-operators-dflvt\" (UID: \"9f328bf3-bacb-4058-bb20-bc99dabaff31\") " pod="openshift-marketplace/redhat-operators-dflvt" Feb 18 15:10:35 crc kubenswrapper[4896]: I0218 15:10:35.949312 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f328bf3-bacb-4058-bb20-bc99dabaff31-utilities\") pod \"redhat-operators-dflvt\" (UID: \"9f328bf3-bacb-4058-bb20-bc99dabaff31\") " pod="openshift-marketplace/redhat-operators-dflvt" Feb 18 15:10:35 crc kubenswrapper[4896]: I0218 15:10:35.949595 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f328bf3-bacb-4058-bb20-bc99dabaff31-catalog-content\") pod \"redhat-operators-dflvt\" (UID: \"9f328bf3-bacb-4058-bb20-bc99dabaff31\") " pod="openshift-marketplace/redhat-operators-dflvt" Feb 18 15:10:35 crc kubenswrapper[4896]: I0218 15:10:35.980667 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4mm2\" (UniqueName: \"kubernetes.io/projected/9f328bf3-bacb-4058-bb20-bc99dabaff31-kube-api-access-h4mm2\") pod \"redhat-operators-dflvt\" (UID: \"9f328bf3-bacb-4058-bb20-bc99dabaff31\") " pod="openshift-marketplace/redhat-operators-dflvt" Feb 18 15:10:36 crc kubenswrapper[4896]: I0218 15:10:36.008551 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dflvt" Feb 18 15:10:36 crc kubenswrapper[4896]: I0218 15:10:36.410100 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dflvt"] Feb 18 15:10:36 crc kubenswrapper[4896]: I0218 15:10:36.534645 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb"] Feb 18 15:10:36 crc kubenswrapper[4896]: I0218 15:10:36.535762 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb" Feb 18 15:10:36 crc kubenswrapper[4896]: I0218 15:10:36.537801 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-9lxwl" Feb 18 15:10:36 crc kubenswrapper[4896]: I0218 15:10:36.547147 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb"] Feb 18 15:10:36 crc kubenswrapper[4896]: I0218 15:10:36.660082 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spb85\" (UniqueName: \"kubernetes.io/projected/688b59fd-cdfb-421e-beff-abf0651ab5f2-kube-api-access-spb85\") pod \"414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb\" (UID: \"688b59fd-cdfb-421e-beff-abf0651ab5f2\") " pod="openstack-operators/414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb" Feb 18 15:10:36 crc kubenswrapper[4896]: I0218 15:10:36.660662 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/688b59fd-cdfb-421e-beff-abf0651ab5f2-bundle\") pod \"414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb\" (UID: \"688b59fd-cdfb-421e-beff-abf0651ab5f2\") " pod="openstack-operators/414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb" Feb 18 15:10:36 crc kubenswrapper[4896]: I0218 15:10:36.660743 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/688b59fd-cdfb-421e-beff-abf0651ab5f2-util\") pod \"414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb\" (UID: \"688b59fd-cdfb-421e-beff-abf0651ab5f2\") " pod="openstack-operators/414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb" Feb 18 15:10:36 crc kubenswrapper[4896]: I0218 15:10:36.761682 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spb85\" (UniqueName: \"kubernetes.io/projected/688b59fd-cdfb-421e-beff-abf0651ab5f2-kube-api-access-spb85\") pod \"414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb\" (UID: \"688b59fd-cdfb-421e-beff-abf0651ab5f2\") " pod="openstack-operators/414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb" Feb 18 15:10:36 crc kubenswrapper[4896]: I0218 15:10:36.761744 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/688b59fd-cdfb-421e-beff-abf0651ab5f2-bundle\") pod \"414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb\" (UID: \"688b59fd-cdfb-421e-beff-abf0651ab5f2\") " pod="openstack-operators/414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb" Feb 18 15:10:36 crc kubenswrapper[4896]: I0218 15:10:36.761792 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/688b59fd-cdfb-421e-beff-abf0651ab5f2-util\") pod \"414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb\" (UID: \"688b59fd-cdfb-421e-beff-abf0651ab5f2\") " pod="openstack-operators/414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb" Feb 18 15:10:36 crc kubenswrapper[4896]: I0218 15:10:36.762280 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/688b59fd-cdfb-421e-beff-abf0651ab5f2-util\") pod \"414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb\" (UID: \"688b59fd-cdfb-421e-beff-abf0651ab5f2\") " pod="openstack-operators/414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb" Feb 18 15:10:36 crc kubenswrapper[4896]: I0218 15:10:36.762512 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/688b59fd-cdfb-421e-beff-abf0651ab5f2-bundle\") pod \"414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb\" (UID: \"688b59fd-cdfb-421e-beff-abf0651ab5f2\") " pod="openstack-operators/414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb" Feb 18 15:10:36 crc kubenswrapper[4896]: I0218 15:10:36.796221 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spb85\" (UniqueName: \"kubernetes.io/projected/688b59fd-cdfb-421e-beff-abf0651ab5f2-kube-api-access-spb85\") pod \"414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb\" (UID: \"688b59fd-cdfb-421e-beff-abf0651ab5f2\") " pod="openstack-operators/414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb" Feb 18 15:10:36 crc kubenswrapper[4896]: I0218 15:10:36.850255 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb" Feb 18 15:10:37 crc kubenswrapper[4896]: I0218 15:10:37.076622 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb"] Feb 18 15:10:37 crc kubenswrapper[4896]: W0218 15:10:37.084608 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod688b59fd_cdfb_421e_beff_abf0651ab5f2.slice/crio-d3f8a05335b038d7df474e2c70d6a1bc794bf962c61f49cfcd14098118ffb55f WatchSource:0}: Error finding container d3f8a05335b038d7df474e2c70d6a1bc794bf962c61f49cfcd14098118ffb55f: Status 404 returned error can't find the container with id d3f8a05335b038d7df474e2c70d6a1bc794bf962c61f49cfcd14098118ffb55f Feb 18 15:10:37 crc kubenswrapper[4896]: I0218 15:10:37.372055 4896 generic.go:334] "Generic (PLEG): container finished" podID="9f328bf3-bacb-4058-bb20-bc99dabaff31" containerID="00a1069b88972b7541065703962ad5e6badbb9e6bc86b8ed9ac500a4b27cc2bf" exitCode=0 Feb 18 15:10:37 crc kubenswrapper[4896]: I0218 15:10:37.372342 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dflvt" event={"ID":"9f328bf3-bacb-4058-bb20-bc99dabaff31","Type":"ContainerDied","Data":"00a1069b88972b7541065703962ad5e6badbb9e6bc86b8ed9ac500a4b27cc2bf"} Feb 18 15:10:37 crc kubenswrapper[4896]: I0218 15:10:37.372524 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dflvt" event={"ID":"9f328bf3-bacb-4058-bb20-bc99dabaff31","Type":"ContainerStarted","Data":"c30f87b0e6e5cd23e0d9ec3a789d10f56a2869b928021d324e962b976f2e57f6"} Feb 18 15:10:37 crc kubenswrapper[4896]: I0218 15:10:37.373954 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb" event={"ID":"688b59fd-cdfb-421e-beff-abf0651ab5f2","Type":"ContainerStarted","Data":"908de0620e5ddcaa5fcabcdea2c387f6996e434c814f1f4cad61691f1582a937"} Feb 18 15:10:37 crc kubenswrapper[4896]: I0218 15:10:37.374018 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb" event={"ID":"688b59fd-cdfb-421e-beff-abf0651ab5f2","Type":"ContainerStarted","Data":"d3f8a05335b038d7df474e2c70d6a1bc794bf962c61f49cfcd14098118ffb55f"} Feb 18 15:10:38 crc kubenswrapper[4896]: I0218 15:10:38.382017 4896 generic.go:334] "Generic (PLEG): container finished" podID="688b59fd-cdfb-421e-beff-abf0651ab5f2" containerID="908de0620e5ddcaa5fcabcdea2c387f6996e434c814f1f4cad61691f1582a937" exitCode=0 Feb 18 15:10:38 crc kubenswrapper[4896]: I0218 15:10:38.382309 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb" event={"ID":"688b59fd-cdfb-421e-beff-abf0651ab5f2","Type":"ContainerDied","Data":"908de0620e5ddcaa5fcabcdea2c387f6996e434c814f1f4cad61691f1582a937"} Feb 18 15:10:38 crc kubenswrapper[4896]: I0218 15:10:38.386939 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dflvt" event={"ID":"9f328bf3-bacb-4058-bb20-bc99dabaff31","Type":"ContainerStarted","Data":"ff040b87961097f6a78477f8123d24f1fbf6845392f8d7f43a848144c0235a8f"} Feb 18 15:10:39 crc kubenswrapper[4896]: I0218 15:10:39.079400 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gz8v8"] Feb 18 15:10:39 crc kubenswrapper[4896]: I0218 15:10:39.080593 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gz8v8" Feb 18 15:10:39 crc kubenswrapper[4896]: I0218 15:10:39.096167 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gz8v8"] Feb 18 15:10:39 crc kubenswrapper[4896]: I0218 15:10:39.194617 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chv9k\" (UniqueName: \"kubernetes.io/projected/11fdd820-f0f9-4250-a424-cd0a0b9c9f29-kube-api-access-chv9k\") pod \"certified-operators-gz8v8\" (UID: \"11fdd820-f0f9-4250-a424-cd0a0b9c9f29\") " pod="openshift-marketplace/certified-operators-gz8v8" Feb 18 15:10:39 crc kubenswrapper[4896]: I0218 15:10:39.194658 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11fdd820-f0f9-4250-a424-cd0a0b9c9f29-catalog-content\") pod \"certified-operators-gz8v8\" (UID: \"11fdd820-f0f9-4250-a424-cd0a0b9c9f29\") " pod="openshift-marketplace/certified-operators-gz8v8" Feb 18 15:10:39 crc kubenswrapper[4896]: I0218 15:10:39.194683 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11fdd820-f0f9-4250-a424-cd0a0b9c9f29-utilities\") pod \"certified-operators-gz8v8\" (UID: \"11fdd820-f0f9-4250-a424-cd0a0b9c9f29\") " pod="openshift-marketplace/certified-operators-gz8v8" Feb 18 15:10:39 crc kubenswrapper[4896]: I0218 15:10:39.295594 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11fdd820-f0f9-4250-a424-cd0a0b9c9f29-catalog-content\") pod \"certified-operators-gz8v8\" (UID: \"11fdd820-f0f9-4250-a424-cd0a0b9c9f29\") " pod="openshift-marketplace/certified-operators-gz8v8" Feb 18 15:10:39 crc kubenswrapper[4896]: I0218 15:10:39.295645 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chv9k\" (UniqueName: \"kubernetes.io/projected/11fdd820-f0f9-4250-a424-cd0a0b9c9f29-kube-api-access-chv9k\") pod \"certified-operators-gz8v8\" (UID: \"11fdd820-f0f9-4250-a424-cd0a0b9c9f29\") " pod="openshift-marketplace/certified-operators-gz8v8" Feb 18 15:10:39 crc kubenswrapper[4896]: I0218 15:10:39.295674 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11fdd820-f0f9-4250-a424-cd0a0b9c9f29-utilities\") pod \"certified-operators-gz8v8\" (UID: \"11fdd820-f0f9-4250-a424-cd0a0b9c9f29\") " pod="openshift-marketplace/certified-operators-gz8v8" Feb 18 15:10:39 crc kubenswrapper[4896]: I0218 15:10:39.296284 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11fdd820-f0f9-4250-a424-cd0a0b9c9f29-utilities\") pod \"certified-operators-gz8v8\" (UID: \"11fdd820-f0f9-4250-a424-cd0a0b9c9f29\") " pod="openshift-marketplace/certified-operators-gz8v8" Feb 18 15:10:39 crc kubenswrapper[4896]: I0218 15:10:39.296554 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11fdd820-f0f9-4250-a424-cd0a0b9c9f29-catalog-content\") pod \"certified-operators-gz8v8\" (UID: \"11fdd820-f0f9-4250-a424-cd0a0b9c9f29\") " pod="openshift-marketplace/certified-operators-gz8v8" Feb 18 15:10:39 crc kubenswrapper[4896]: I0218 15:10:39.321879 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chv9k\" (UniqueName: \"kubernetes.io/projected/11fdd820-f0f9-4250-a424-cd0a0b9c9f29-kube-api-access-chv9k\") pod \"certified-operators-gz8v8\" (UID: \"11fdd820-f0f9-4250-a424-cd0a0b9c9f29\") " pod="openshift-marketplace/certified-operators-gz8v8" Feb 18 15:10:39 crc kubenswrapper[4896]: I0218 15:10:39.444209 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gz8v8" Feb 18 15:10:39 crc kubenswrapper[4896]: I0218 15:10:39.718495 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gz8v8"] Feb 18 15:10:39 crc kubenswrapper[4896]: W0218 15:10:39.722867 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod11fdd820_f0f9_4250_a424_cd0a0b9c9f29.slice/crio-3b24b4672d8f947d773418836237d6f458a170f9b52f4b40624b9e22d0bda8cc WatchSource:0}: Error finding container 3b24b4672d8f947d773418836237d6f458a170f9b52f4b40624b9e22d0bda8cc: Status 404 returned error can't find the container with id 3b24b4672d8f947d773418836237d6f458a170f9b52f4b40624b9e22d0bda8cc Feb 18 15:10:40 crc kubenswrapper[4896]: I0218 15:10:40.399452 4896 generic.go:334] "Generic (PLEG): container finished" podID="9f328bf3-bacb-4058-bb20-bc99dabaff31" containerID="ff040b87961097f6a78477f8123d24f1fbf6845392f8d7f43a848144c0235a8f" exitCode=0 Feb 18 15:10:40 crc kubenswrapper[4896]: I0218 15:10:40.399502 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dflvt" event={"ID":"9f328bf3-bacb-4058-bb20-bc99dabaff31","Type":"ContainerDied","Data":"ff040b87961097f6a78477f8123d24f1fbf6845392f8d7f43a848144c0235a8f"} Feb 18 15:10:40 crc kubenswrapper[4896]: I0218 15:10:40.402345 4896 generic.go:334] "Generic (PLEG): container finished" podID="688b59fd-cdfb-421e-beff-abf0651ab5f2" containerID="545a4c3038c10b56cb072ea7a77dc1ea0326498db46bc2deb1b82890abab8d23" exitCode=0 Feb 18 15:10:40 crc kubenswrapper[4896]: I0218 15:10:40.402633 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb" event={"ID":"688b59fd-cdfb-421e-beff-abf0651ab5f2","Type":"ContainerDied","Data":"545a4c3038c10b56cb072ea7a77dc1ea0326498db46bc2deb1b82890abab8d23"} Feb 18 15:10:40 crc kubenswrapper[4896]: I0218 15:10:40.406518 4896 generic.go:334] "Generic (PLEG): container finished" podID="11fdd820-f0f9-4250-a424-cd0a0b9c9f29" containerID="2b099671ad1a18101b16a5491a85a0a06dba818493b09f89741a37bbd2804560" exitCode=0 Feb 18 15:10:40 crc kubenswrapper[4896]: I0218 15:10:40.406570 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gz8v8" event={"ID":"11fdd820-f0f9-4250-a424-cd0a0b9c9f29","Type":"ContainerDied","Data":"2b099671ad1a18101b16a5491a85a0a06dba818493b09f89741a37bbd2804560"} Feb 18 15:10:40 crc kubenswrapper[4896]: I0218 15:10:40.406607 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gz8v8" event={"ID":"11fdd820-f0f9-4250-a424-cd0a0b9c9f29","Type":"ContainerStarted","Data":"3b24b4672d8f947d773418836237d6f458a170f9b52f4b40624b9e22d0bda8cc"} Feb 18 15:10:41 crc kubenswrapper[4896]: I0218 15:10:41.414505 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gz8v8" event={"ID":"11fdd820-f0f9-4250-a424-cd0a0b9c9f29","Type":"ContainerStarted","Data":"b8c0d4c48773ffdd100448267fd41c86140c515754f2086382fa97cbed3a5cb6"} Feb 18 15:10:41 crc kubenswrapper[4896]: I0218 15:10:41.416544 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dflvt" event={"ID":"9f328bf3-bacb-4058-bb20-bc99dabaff31","Type":"ContainerStarted","Data":"439e6446916aa08472620981e51299474c705bcd7ee0edd555ca02c9d2b791fe"} Feb 18 15:10:41 crc kubenswrapper[4896]: I0218 15:10:41.419346 4896 generic.go:334] "Generic (PLEG): container finished" podID="688b59fd-cdfb-421e-beff-abf0651ab5f2" containerID="0687739882efffc11529c164da12c3b147f13de24812ead0f6409c7e7409f7de" exitCode=0 Feb 18 15:10:41 crc kubenswrapper[4896]: I0218 15:10:41.419377 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb" event={"ID":"688b59fd-cdfb-421e-beff-abf0651ab5f2","Type":"ContainerDied","Data":"0687739882efffc11529c164da12c3b147f13de24812ead0f6409c7e7409f7de"} Feb 18 15:10:41 crc kubenswrapper[4896]: I0218 15:10:41.472606 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dflvt" podStartSLOduration=3.068204564 podStartE2EDuration="6.472588899s" podCreationTimestamp="2026-02-18 15:10:35 +0000 UTC" firstStartedPulling="2026-02-18 15:10:37.373685568 +0000 UTC m=+936.303019771" lastFinishedPulling="2026-02-18 15:10:40.778069893 +0000 UTC m=+939.707404106" observedRunningTime="2026-02-18 15:10:41.469427045 +0000 UTC m=+940.398761258" watchObservedRunningTime="2026-02-18 15:10:41.472588899 +0000 UTC m=+940.401923112" Feb 18 15:10:42 crc kubenswrapper[4896]: I0218 15:10:42.428354 4896 generic.go:334] "Generic (PLEG): container finished" podID="11fdd820-f0f9-4250-a424-cd0a0b9c9f29" containerID="b8c0d4c48773ffdd100448267fd41c86140c515754f2086382fa97cbed3a5cb6" exitCode=0 Feb 18 15:10:42 crc kubenswrapper[4896]: I0218 15:10:42.428453 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gz8v8" event={"ID":"11fdd820-f0f9-4250-a424-cd0a0b9c9f29","Type":"ContainerDied","Data":"b8c0d4c48773ffdd100448267fd41c86140c515754f2086382fa97cbed3a5cb6"} Feb 18 15:10:42 crc kubenswrapper[4896]: I0218 15:10:42.756964 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb" Feb 18 15:10:42 crc kubenswrapper[4896]: I0218 15:10:42.840043 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/688b59fd-cdfb-421e-beff-abf0651ab5f2-util\") pod \"688b59fd-cdfb-421e-beff-abf0651ab5f2\" (UID: \"688b59fd-cdfb-421e-beff-abf0651ab5f2\") " Feb 18 15:10:42 crc kubenswrapper[4896]: I0218 15:10:42.840120 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spb85\" (UniqueName: \"kubernetes.io/projected/688b59fd-cdfb-421e-beff-abf0651ab5f2-kube-api-access-spb85\") pod \"688b59fd-cdfb-421e-beff-abf0651ab5f2\" (UID: \"688b59fd-cdfb-421e-beff-abf0651ab5f2\") " Feb 18 15:10:42 crc kubenswrapper[4896]: I0218 15:10:42.840200 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/688b59fd-cdfb-421e-beff-abf0651ab5f2-bundle\") pod \"688b59fd-cdfb-421e-beff-abf0651ab5f2\" (UID: \"688b59fd-cdfb-421e-beff-abf0651ab5f2\") " Feb 18 15:10:42 crc kubenswrapper[4896]: I0218 15:10:42.841100 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/688b59fd-cdfb-421e-beff-abf0651ab5f2-bundle" (OuterVolumeSpecName: "bundle") pod "688b59fd-cdfb-421e-beff-abf0651ab5f2" (UID: "688b59fd-cdfb-421e-beff-abf0651ab5f2"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:10:42 crc kubenswrapper[4896]: I0218 15:10:42.846841 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/688b59fd-cdfb-421e-beff-abf0651ab5f2-kube-api-access-spb85" (OuterVolumeSpecName: "kube-api-access-spb85") pod "688b59fd-cdfb-421e-beff-abf0651ab5f2" (UID: "688b59fd-cdfb-421e-beff-abf0651ab5f2"). InnerVolumeSpecName "kube-api-access-spb85". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:10:42 crc kubenswrapper[4896]: I0218 15:10:42.941848 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spb85\" (UniqueName: \"kubernetes.io/projected/688b59fd-cdfb-421e-beff-abf0651ab5f2-kube-api-access-spb85\") on node \"crc\" DevicePath \"\"" Feb 18 15:10:42 crc kubenswrapper[4896]: I0218 15:10:42.941910 4896 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/688b59fd-cdfb-421e-beff-abf0651ab5f2-bundle\") on node \"crc\" DevicePath \"\"" Feb 18 15:10:43 crc kubenswrapper[4896]: I0218 15:10:43.181262 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/688b59fd-cdfb-421e-beff-abf0651ab5f2-util" (OuterVolumeSpecName: "util") pod "688b59fd-cdfb-421e-beff-abf0651ab5f2" (UID: "688b59fd-cdfb-421e-beff-abf0651ab5f2"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:10:43 crc kubenswrapper[4896]: I0218 15:10:43.247068 4896 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/688b59fd-cdfb-421e-beff-abf0651ab5f2-util\") on node \"crc\" DevicePath \"\"" Feb 18 15:10:43 crc kubenswrapper[4896]: I0218 15:10:43.437374 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb" event={"ID":"688b59fd-cdfb-421e-beff-abf0651ab5f2","Type":"ContainerDied","Data":"d3f8a05335b038d7df474e2c70d6a1bc794bf962c61f49cfcd14098118ffb55f"} Feb 18 15:10:43 crc kubenswrapper[4896]: I0218 15:10:43.437417 4896 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3f8a05335b038d7df474e2c70d6a1bc794bf962c61f49cfcd14098118ffb55f" Feb 18 15:10:43 crc kubenswrapper[4896]: I0218 15:10:43.437390 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb" Feb 18 15:10:43 crc kubenswrapper[4896]: I0218 15:10:43.440027 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gz8v8" event={"ID":"11fdd820-f0f9-4250-a424-cd0a0b9c9f29","Type":"ContainerStarted","Data":"15bfe883063030fecc8111b81458d2c18908435577c211795665f72616b17e31"} Feb 18 15:10:43 crc kubenswrapper[4896]: I0218 15:10:43.456731 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gz8v8" podStartSLOduration=1.918315408 podStartE2EDuration="4.456716555s" podCreationTimestamp="2026-02-18 15:10:39 +0000 UTC" firstStartedPulling="2026-02-18 15:10:40.408105243 +0000 UTC m=+939.337439456" lastFinishedPulling="2026-02-18 15:10:42.94650639 +0000 UTC m=+941.875840603" observedRunningTime="2026-02-18 15:10:43.454136566 +0000 UTC m=+942.383470779" watchObservedRunningTime="2026-02-18 15:10:43.456716555 +0000 UTC m=+942.386050768" Feb 18 15:10:46 crc kubenswrapper[4896]: I0218 15:10:46.009254 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dflvt" Feb 18 15:10:46 crc kubenswrapper[4896]: I0218 15:10:46.009599 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dflvt" Feb 18 15:10:46 crc kubenswrapper[4896]: I0218 15:10:46.464030 4896 generic.go:334] "Generic (PLEG): container finished" podID="c5ce908c-bcd7-456e-9ab2-abe730180641" containerID="964908cec1ea5b2b53e165375d511f323d804bf9fc89abc3a0d6d01d9b60bbb3" exitCode=0 Feb 18 15:10:46 crc kubenswrapper[4896]: I0218 15:10:46.464077 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/rabbitmq-server-0" event={"ID":"c5ce908c-bcd7-456e-9ab2-abe730180641","Type":"ContainerDied","Data":"964908cec1ea5b2b53e165375d511f323d804bf9fc89abc3a0d6d01d9b60bbb3"} Feb 18 15:10:47 crc kubenswrapper[4896]: I0218 15:10:47.049576 4896 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dflvt" podUID="9f328bf3-bacb-4058-bb20-bc99dabaff31" containerName="registry-server" probeResult="failure" output=< Feb 18 15:10:47 crc kubenswrapper[4896]: timeout: failed to connect service ":50051" within 1s Feb 18 15:10:47 crc kubenswrapper[4896]: > Feb 18 15:10:47 crc kubenswrapper[4896]: I0218 15:10:47.471662 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/rabbitmq-server-0" event={"ID":"c5ce908c-bcd7-456e-9ab2-abe730180641","Type":"ContainerStarted","Data":"dca4a7fd841538889dee779d1d10684bcd7ff61b265e84cb5a235186cb1328e6"} Feb 18 15:10:47 crc kubenswrapper[4896]: I0218 15:10:47.471893 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="test-operator-kuttl-tests/rabbitmq-server-0" Feb 18 15:10:47 crc kubenswrapper[4896]: I0218 15:10:47.495912 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="test-operator-kuttl-tests/rabbitmq-server-0" podStartSLOduration=36.144604606 podStartE2EDuration="44.495895173s" podCreationTimestamp="2026-02-18 15:10:03 +0000 UTC" firstStartedPulling="2026-02-18 15:10:04.818670628 +0000 UTC m=+903.748004841" lastFinishedPulling="2026-02-18 15:10:13.169961195 +0000 UTC m=+912.099295408" observedRunningTime="2026-02-18 15:10:47.495502112 +0000 UTC m=+946.424836335" watchObservedRunningTime="2026-02-18 15:10:47.495895173 +0000 UTC m=+946.425229386" Feb 18 15:10:49 crc kubenswrapper[4896]: I0218 15:10:49.444371 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gz8v8" Feb 18 15:10:49 crc kubenswrapper[4896]: I0218 15:10:49.445474 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gz8v8" Feb 18 15:10:49 crc kubenswrapper[4896]: I0218 15:10:49.485243 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gz8v8" Feb 18 15:10:50 crc kubenswrapper[4896]: I0218 15:10:50.529411 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gz8v8" Feb 18 15:10:52 crc kubenswrapper[4896]: I0218 15:10:52.872506 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gz8v8"] Feb 18 15:10:52 crc kubenswrapper[4896]: I0218 15:10:52.873035 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gz8v8" podUID="11fdd820-f0f9-4250-a424-cd0a0b9c9f29" containerName="registry-server" containerID="cri-o://15bfe883063030fecc8111b81458d2c18908435577c211795665f72616b17e31" gracePeriod=2 Feb 18 15:10:53 crc kubenswrapper[4896]: I0218 15:10:53.276927 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gz8v8" Feb 18 15:10:53 crc kubenswrapper[4896]: I0218 15:10:53.295499 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chv9k\" (UniqueName: \"kubernetes.io/projected/11fdd820-f0f9-4250-a424-cd0a0b9c9f29-kube-api-access-chv9k\") pod \"11fdd820-f0f9-4250-a424-cd0a0b9c9f29\" (UID: \"11fdd820-f0f9-4250-a424-cd0a0b9c9f29\") " Feb 18 15:10:53 crc kubenswrapper[4896]: I0218 15:10:53.296403 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11fdd820-f0f9-4250-a424-cd0a0b9c9f29-catalog-content\") pod \"11fdd820-f0f9-4250-a424-cd0a0b9c9f29\" (UID: \"11fdd820-f0f9-4250-a424-cd0a0b9c9f29\") " Feb 18 15:10:53 crc kubenswrapper[4896]: I0218 15:10:53.296487 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11fdd820-f0f9-4250-a424-cd0a0b9c9f29-utilities\") pod \"11fdd820-f0f9-4250-a424-cd0a0b9c9f29\" (UID: \"11fdd820-f0f9-4250-a424-cd0a0b9c9f29\") " Feb 18 15:10:53 crc kubenswrapper[4896]: I0218 15:10:53.297523 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11fdd820-f0f9-4250-a424-cd0a0b9c9f29-utilities" (OuterVolumeSpecName: "utilities") pod "11fdd820-f0f9-4250-a424-cd0a0b9c9f29" (UID: "11fdd820-f0f9-4250-a424-cd0a0b9c9f29"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:10:53 crc kubenswrapper[4896]: I0218 15:10:53.301598 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11fdd820-f0f9-4250-a424-cd0a0b9c9f29-kube-api-access-chv9k" (OuterVolumeSpecName: "kube-api-access-chv9k") pod "11fdd820-f0f9-4250-a424-cd0a0b9c9f29" (UID: "11fdd820-f0f9-4250-a424-cd0a0b9c9f29"). InnerVolumeSpecName "kube-api-access-chv9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:10:53 crc kubenswrapper[4896]: I0218 15:10:53.357169 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11fdd820-f0f9-4250-a424-cd0a0b9c9f29-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "11fdd820-f0f9-4250-a424-cd0a0b9c9f29" (UID: "11fdd820-f0f9-4250-a424-cd0a0b9c9f29"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:10:53 crc kubenswrapper[4896]: I0218 15:10:53.397990 4896 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11fdd820-f0f9-4250-a424-cd0a0b9c9f29-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 15:10:53 crc kubenswrapper[4896]: I0218 15:10:53.398030 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chv9k\" (UniqueName: \"kubernetes.io/projected/11fdd820-f0f9-4250-a424-cd0a0b9c9f29-kube-api-access-chv9k\") on node \"crc\" DevicePath \"\"" Feb 18 15:10:53 crc kubenswrapper[4896]: I0218 15:10:53.398042 4896 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11fdd820-f0f9-4250-a424-cd0a0b9c9f29-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 15:10:53 crc kubenswrapper[4896]: I0218 15:10:53.513569 4896 generic.go:334] "Generic (PLEG): container finished" podID="11fdd820-f0f9-4250-a424-cd0a0b9c9f29" containerID="15bfe883063030fecc8111b81458d2c18908435577c211795665f72616b17e31" exitCode=0 Feb 18 15:10:53 crc kubenswrapper[4896]: I0218 15:10:53.513618 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gz8v8" event={"ID":"11fdd820-f0f9-4250-a424-cd0a0b9c9f29","Type":"ContainerDied","Data":"15bfe883063030fecc8111b81458d2c18908435577c211795665f72616b17e31"} Feb 18 15:10:53 crc kubenswrapper[4896]: I0218 15:10:53.513648 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gz8v8" event={"ID":"11fdd820-f0f9-4250-a424-cd0a0b9c9f29","Type":"ContainerDied","Data":"3b24b4672d8f947d773418836237d6f458a170f9b52f4b40624b9e22d0bda8cc"} Feb 18 15:10:53 crc kubenswrapper[4896]: I0218 15:10:53.513670 4896 scope.go:117] "RemoveContainer" containerID="15bfe883063030fecc8111b81458d2c18908435577c211795665f72616b17e31" Feb 18 15:10:53 crc kubenswrapper[4896]: I0218 15:10:53.513766 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gz8v8" Feb 18 15:10:53 crc kubenswrapper[4896]: I0218 15:10:53.531984 4896 scope.go:117] "RemoveContainer" containerID="b8c0d4c48773ffdd100448267fd41c86140c515754f2086382fa97cbed3a5cb6" Feb 18 15:10:53 crc kubenswrapper[4896]: I0218 15:10:53.556908 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gz8v8"] Feb 18 15:10:53 crc kubenswrapper[4896]: I0218 15:10:53.562738 4896 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gz8v8"] Feb 18 15:10:53 crc kubenswrapper[4896]: I0218 15:10:53.565524 4896 scope.go:117] "RemoveContainer" containerID="2b099671ad1a18101b16a5491a85a0a06dba818493b09f89741a37bbd2804560" Feb 18 15:10:53 crc kubenswrapper[4896]: I0218 15:10:53.580878 4896 scope.go:117] "RemoveContainer" containerID="15bfe883063030fecc8111b81458d2c18908435577c211795665f72616b17e31" Feb 18 15:10:53 crc kubenswrapper[4896]: E0218 15:10:53.581428 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15bfe883063030fecc8111b81458d2c18908435577c211795665f72616b17e31\": container with ID starting with 15bfe883063030fecc8111b81458d2c18908435577c211795665f72616b17e31 not found: ID does not exist" containerID="15bfe883063030fecc8111b81458d2c18908435577c211795665f72616b17e31" Feb 18 15:10:53 crc kubenswrapper[4896]: I0218 15:10:53.581480 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15bfe883063030fecc8111b81458d2c18908435577c211795665f72616b17e31"} err="failed to get container status \"15bfe883063030fecc8111b81458d2c18908435577c211795665f72616b17e31\": rpc error: code = NotFound desc = could not find container \"15bfe883063030fecc8111b81458d2c18908435577c211795665f72616b17e31\": container with ID starting with 15bfe883063030fecc8111b81458d2c18908435577c211795665f72616b17e31 not found: ID does not exist" Feb 18 15:10:53 crc kubenswrapper[4896]: I0218 15:10:53.581562 4896 scope.go:117] "RemoveContainer" containerID="b8c0d4c48773ffdd100448267fd41c86140c515754f2086382fa97cbed3a5cb6" Feb 18 15:10:53 crc kubenswrapper[4896]: E0218 15:10:53.581907 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8c0d4c48773ffdd100448267fd41c86140c515754f2086382fa97cbed3a5cb6\": container with ID starting with b8c0d4c48773ffdd100448267fd41c86140c515754f2086382fa97cbed3a5cb6 not found: ID does not exist" containerID="b8c0d4c48773ffdd100448267fd41c86140c515754f2086382fa97cbed3a5cb6" Feb 18 15:10:53 crc kubenswrapper[4896]: I0218 15:10:53.581977 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8c0d4c48773ffdd100448267fd41c86140c515754f2086382fa97cbed3a5cb6"} err="failed to get container status \"b8c0d4c48773ffdd100448267fd41c86140c515754f2086382fa97cbed3a5cb6\": rpc error: code = NotFound desc = could not find container \"b8c0d4c48773ffdd100448267fd41c86140c515754f2086382fa97cbed3a5cb6\": container with ID starting with b8c0d4c48773ffdd100448267fd41c86140c515754f2086382fa97cbed3a5cb6 not found: ID does not exist" Feb 18 15:10:53 crc kubenswrapper[4896]: I0218 15:10:53.582011 4896 scope.go:117] "RemoveContainer" containerID="2b099671ad1a18101b16a5491a85a0a06dba818493b09f89741a37bbd2804560" Feb 18 15:10:53 crc kubenswrapper[4896]: E0218 15:10:53.582317 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b099671ad1a18101b16a5491a85a0a06dba818493b09f89741a37bbd2804560\": container with ID starting with 2b099671ad1a18101b16a5491a85a0a06dba818493b09f89741a37bbd2804560 not found: ID does not exist" containerID="2b099671ad1a18101b16a5491a85a0a06dba818493b09f89741a37bbd2804560" Feb 18 15:10:53 crc kubenswrapper[4896]: I0218 15:10:53.582342 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b099671ad1a18101b16a5491a85a0a06dba818493b09f89741a37bbd2804560"} err="failed to get container status \"2b099671ad1a18101b16a5491a85a0a06dba818493b09f89741a37bbd2804560\": rpc error: code = NotFound desc = could not find container \"2b099671ad1a18101b16a5491a85a0a06dba818493b09f89741a37bbd2804560\": container with ID starting with 2b099671ad1a18101b16a5491a85a0a06dba818493b09f89741a37bbd2804560 not found: ID does not exist" Feb 18 15:10:53 crc kubenswrapper[4896]: I0218 15:10:53.936748 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11fdd820-f0f9-4250-a424-cd0a0b9c9f29" path="/var/lib/kubelet/pods/11fdd820-f0f9-4250-a424-cd0a0b9c9f29/volumes" Feb 18 15:10:54 crc kubenswrapper[4896]: I0218 15:10:54.350357 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-77bdc966f6-nnk54"] Feb 18 15:10:54 crc kubenswrapper[4896]: E0218 15:10:54.350868 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="688b59fd-cdfb-421e-beff-abf0651ab5f2" containerName="pull" Feb 18 15:10:54 crc kubenswrapper[4896]: I0218 15:10:54.350885 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="688b59fd-cdfb-421e-beff-abf0651ab5f2" containerName="pull" Feb 18 15:10:54 crc kubenswrapper[4896]: E0218 15:10:54.350928 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11fdd820-f0f9-4250-a424-cd0a0b9c9f29" containerName="extract-content" Feb 18 15:10:54 crc kubenswrapper[4896]: I0218 15:10:54.350934 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="11fdd820-f0f9-4250-a424-cd0a0b9c9f29" containerName="extract-content" Feb 18 15:10:54 crc kubenswrapper[4896]: E0218 15:10:54.350945 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11fdd820-f0f9-4250-a424-cd0a0b9c9f29" containerName="extract-utilities" Feb 18 15:10:54 crc kubenswrapper[4896]: I0218 15:10:54.350951 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="11fdd820-f0f9-4250-a424-cd0a0b9c9f29" containerName="extract-utilities" Feb 18 15:10:54 crc kubenswrapper[4896]: E0218 15:10:54.350964 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="688b59fd-cdfb-421e-beff-abf0651ab5f2" containerName="util" Feb 18 15:10:54 crc kubenswrapper[4896]: I0218 15:10:54.350969 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="688b59fd-cdfb-421e-beff-abf0651ab5f2" containerName="util" Feb 18 15:10:54 crc kubenswrapper[4896]: E0218 15:10:54.350977 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="688b59fd-cdfb-421e-beff-abf0651ab5f2" containerName="extract" Feb 18 15:10:54 crc kubenswrapper[4896]: I0218 15:10:54.350982 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="688b59fd-cdfb-421e-beff-abf0651ab5f2" containerName="extract" Feb 18 15:10:54 crc kubenswrapper[4896]: E0218 15:10:54.350995 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11fdd820-f0f9-4250-a424-cd0a0b9c9f29" containerName="registry-server" Feb 18 15:10:54 crc kubenswrapper[4896]: I0218 15:10:54.351002 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="11fdd820-f0f9-4250-a424-cd0a0b9c9f29" containerName="registry-server" Feb 18 15:10:54 crc kubenswrapper[4896]: I0218 15:10:54.351104 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="688b59fd-cdfb-421e-beff-abf0651ab5f2" containerName="extract" Feb 18 15:10:54 crc kubenswrapper[4896]: I0218 15:10:54.351123 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="11fdd820-f0f9-4250-a424-cd0a0b9c9f29" containerName="registry-server" Feb 18 15:10:54 crc kubenswrapper[4896]: I0218 15:10:54.351543 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-77bdc966f6-nnk54" Feb 18 15:10:54 crc kubenswrapper[4896]: I0218 15:10:54.353832 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-77hgl" Feb 18 15:10:54 crc kubenswrapper[4896]: I0218 15:10:54.354258 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-service-cert" Feb 18 15:10:54 crc kubenswrapper[4896]: I0218 15:10:54.366315 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-77bdc966f6-nnk54"] Feb 18 15:10:54 crc kubenswrapper[4896]: I0218 15:10:54.410652 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/71d66b07-2d37-4b32-9d3e-6953073d0697-webhook-cert\") pod \"keystone-operator-controller-manager-77bdc966f6-nnk54\" (UID: \"71d66b07-2d37-4b32-9d3e-6953073d0697\") " pod="openstack-operators/keystone-operator-controller-manager-77bdc966f6-nnk54" Feb 18 15:10:54 crc kubenswrapper[4896]: I0218 15:10:54.410742 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/71d66b07-2d37-4b32-9d3e-6953073d0697-apiservice-cert\") pod \"keystone-operator-controller-manager-77bdc966f6-nnk54\" (UID: \"71d66b07-2d37-4b32-9d3e-6953073d0697\") " pod="openstack-operators/keystone-operator-controller-manager-77bdc966f6-nnk54" Feb 18 15:10:54 crc kubenswrapper[4896]: I0218 15:10:54.410768 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcfd6\" (UniqueName: \"kubernetes.io/projected/71d66b07-2d37-4b32-9d3e-6953073d0697-kube-api-access-pcfd6\") pod \"keystone-operator-controller-manager-77bdc966f6-nnk54\" (UID: \"71d66b07-2d37-4b32-9d3e-6953073d0697\") " pod="openstack-operators/keystone-operator-controller-manager-77bdc966f6-nnk54" Feb 18 15:10:54 crc kubenswrapper[4896]: I0218 15:10:54.511858 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/71d66b07-2d37-4b32-9d3e-6953073d0697-apiservice-cert\") pod \"keystone-operator-controller-manager-77bdc966f6-nnk54\" (UID: \"71d66b07-2d37-4b32-9d3e-6953073d0697\") " pod="openstack-operators/keystone-operator-controller-manager-77bdc966f6-nnk54" Feb 18 15:10:54 crc kubenswrapper[4896]: I0218 15:10:54.511899 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcfd6\" (UniqueName: \"kubernetes.io/projected/71d66b07-2d37-4b32-9d3e-6953073d0697-kube-api-access-pcfd6\") pod \"keystone-operator-controller-manager-77bdc966f6-nnk54\" (UID: \"71d66b07-2d37-4b32-9d3e-6953073d0697\") " pod="openstack-operators/keystone-operator-controller-manager-77bdc966f6-nnk54" Feb 18 15:10:54 crc kubenswrapper[4896]: I0218 15:10:54.511956 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/71d66b07-2d37-4b32-9d3e-6953073d0697-webhook-cert\") pod \"keystone-operator-controller-manager-77bdc966f6-nnk54\" (UID: \"71d66b07-2d37-4b32-9d3e-6953073d0697\") " pod="openstack-operators/keystone-operator-controller-manager-77bdc966f6-nnk54" Feb 18 15:10:54 crc kubenswrapper[4896]: I0218 15:10:54.517431 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/71d66b07-2d37-4b32-9d3e-6953073d0697-webhook-cert\") pod \"keystone-operator-controller-manager-77bdc966f6-nnk54\" (UID: \"71d66b07-2d37-4b32-9d3e-6953073d0697\") " pod="openstack-operators/keystone-operator-controller-manager-77bdc966f6-nnk54" Feb 18 15:10:54 crc kubenswrapper[4896]: I0218 15:10:54.518368 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/71d66b07-2d37-4b32-9d3e-6953073d0697-apiservice-cert\") pod \"keystone-operator-controller-manager-77bdc966f6-nnk54\" (UID: \"71d66b07-2d37-4b32-9d3e-6953073d0697\") " pod="openstack-operators/keystone-operator-controller-manager-77bdc966f6-nnk54" Feb 18 15:10:54 crc kubenswrapper[4896]: I0218 15:10:54.528847 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcfd6\" (UniqueName: \"kubernetes.io/projected/71d66b07-2d37-4b32-9d3e-6953073d0697-kube-api-access-pcfd6\") pod \"keystone-operator-controller-manager-77bdc966f6-nnk54\" (UID: \"71d66b07-2d37-4b32-9d3e-6953073d0697\") " pod="openstack-operators/keystone-operator-controller-manager-77bdc966f6-nnk54" Feb 18 15:10:54 crc kubenswrapper[4896]: I0218 15:10:54.668133 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-77bdc966f6-nnk54" Feb 18 15:10:55 crc kubenswrapper[4896]: I0218 15:10:55.071794 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-77bdc966f6-nnk54"] Feb 18 15:10:55 crc kubenswrapper[4896]: W0218 15:10:55.075073 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71d66b07_2d37_4b32_9d3e_6953073d0697.slice/crio-bcfa62a21af5c37468f6be1814f2cc2a0db5b148d6be24449e49bcb4c87de837 WatchSource:0}: Error finding container bcfa62a21af5c37468f6be1814f2cc2a0db5b148d6be24449e49bcb4c87de837: Status 404 returned error can't find the container with id bcfa62a21af5c37468f6be1814f2cc2a0db5b148d6be24449e49bcb4c87de837 Feb 18 15:10:55 crc kubenswrapper[4896]: I0218 15:10:55.534036 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-77bdc966f6-nnk54" event={"ID":"71d66b07-2d37-4b32-9d3e-6953073d0697","Type":"ContainerStarted","Data":"bcfa62a21af5c37468f6be1814f2cc2a0db5b148d6be24449e49bcb4c87de837"} Feb 18 15:10:56 crc kubenswrapper[4896]: I0218 15:10:56.050629 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dflvt" Feb 18 15:10:56 crc kubenswrapper[4896]: I0218 15:10:56.095311 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dflvt" Feb 18 15:10:58 crc kubenswrapper[4896]: I0218 15:10:58.555131 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-77bdc966f6-nnk54" event={"ID":"71d66b07-2d37-4b32-9d3e-6953073d0697","Type":"ContainerStarted","Data":"7b755ad6e4c056248f324b06ae6d3c6649f46836a34f49e8dd061b07f704fb46"} Feb 18 15:10:58 crc kubenswrapper[4896]: I0218 15:10:58.555459 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-77bdc966f6-nnk54" Feb 18 15:10:58 crc kubenswrapper[4896]: I0218 15:10:58.575909 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-77bdc966f6-nnk54" podStartSLOduration=1.486402193 podStartE2EDuration="4.575889367s" podCreationTimestamp="2026-02-18 15:10:54 +0000 UTC" firstStartedPulling="2026-02-18 15:10:55.077912703 +0000 UTC m=+954.007246916" lastFinishedPulling="2026-02-18 15:10:58.167399877 +0000 UTC m=+957.096734090" observedRunningTime="2026-02-18 15:10:58.574460699 +0000 UTC m=+957.503794932" watchObservedRunningTime="2026-02-18 15:10:58.575889367 +0000 UTC m=+957.505223580" Feb 18 15:11:00 crc kubenswrapper[4896]: I0218 15:11:00.673590 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dflvt"] Feb 18 15:11:00 crc kubenswrapper[4896]: I0218 15:11:00.674120 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dflvt" podUID="9f328bf3-bacb-4058-bb20-bc99dabaff31" containerName="registry-server" containerID="cri-o://439e6446916aa08472620981e51299474c705bcd7ee0edd555ca02c9d2b791fe" gracePeriod=2 Feb 18 15:11:01 crc kubenswrapper[4896]: I0218 15:11:01.066734 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dflvt" Feb 18 15:11:01 crc kubenswrapper[4896]: I0218 15:11:01.213677 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f328bf3-bacb-4058-bb20-bc99dabaff31-catalog-content\") pod \"9f328bf3-bacb-4058-bb20-bc99dabaff31\" (UID: \"9f328bf3-bacb-4058-bb20-bc99dabaff31\") " Feb 18 15:11:01 crc kubenswrapper[4896]: I0218 15:11:01.213768 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4mm2\" (UniqueName: \"kubernetes.io/projected/9f328bf3-bacb-4058-bb20-bc99dabaff31-kube-api-access-h4mm2\") pod \"9f328bf3-bacb-4058-bb20-bc99dabaff31\" (UID: \"9f328bf3-bacb-4058-bb20-bc99dabaff31\") " Feb 18 15:11:01 crc kubenswrapper[4896]: I0218 15:11:01.213817 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f328bf3-bacb-4058-bb20-bc99dabaff31-utilities\") pod \"9f328bf3-bacb-4058-bb20-bc99dabaff31\" (UID: \"9f328bf3-bacb-4058-bb20-bc99dabaff31\") " Feb 18 15:11:01 crc kubenswrapper[4896]: I0218 15:11:01.214714 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f328bf3-bacb-4058-bb20-bc99dabaff31-utilities" (OuterVolumeSpecName: "utilities") pod "9f328bf3-bacb-4058-bb20-bc99dabaff31" (UID: "9f328bf3-bacb-4058-bb20-bc99dabaff31"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:11:01 crc kubenswrapper[4896]: I0218 15:11:01.219369 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f328bf3-bacb-4058-bb20-bc99dabaff31-kube-api-access-h4mm2" (OuterVolumeSpecName: "kube-api-access-h4mm2") pod "9f328bf3-bacb-4058-bb20-bc99dabaff31" (UID: "9f328bf3-bacb-4058-bb20-bc99dabaff31"). InnerVolumeSpecName "kube-api-access-h4mm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:11:01 crc kubenswrapper[4896]: I0218 15:11:01.315616 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4mm2\" (UniqueName: \"kubernetes.io/projected/9f328bf3-bacb-4058-bb20-bc99dabaff31-kube-api-access-h4mm2\") on node \"crc\" DevicePath \"\"" Feb 18 15:11:01 crc kubenswrapper[4896]: I0218 15:11:01.315652 4896 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f328bf3-bacb-4058-bb20-bc99dabaff31-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 15:11:01 crc kubenswrapper[4896]: I0218 15:11:01.332525 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f328bf3-bacb-4058-bb20-bc99dabaff31-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9f328bf3-bacb-4058-bb20-bc99dabaff31" (UID: "9f328bf3-bacb-4058-bb20-bc99dabaff31"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:11:01 crc kubenswrapper[4896]: I0218 15:11:01.416753 4896 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f328bf3-bacb-4058-bb20-bc99dabaff31-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 15:11:01 crc kubenswrapper[4896]: I0218 15:11:01.574660 4896 generic.go:334] "Generic (PLEG): container finished" podID="9f328bf3-bacb-4058-bb20-bc99dabaff31" containerID="439e6446916aa08472620981e51299474c705bcd7ee0edd555ca02c9d2b791fe" exitCode=0 Feb 18 15:11:01 crc kubenswrapper[4896]: I0218 15:11:01.574720 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dflvt" event={"ID":"9f328bf3-bacb-4058-bb20-bc99dabaff31","Type":"ContainerDied","Data":"439e6446916aa08472620981e51299474c705bcd7ee0edd555ca02c9d2b791fe"} Feb 18 15:11:01 crc kubenswrapper[4896]: I0218 15:11:01.574752 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dflvt" event={"ID":"9f328bf3-bacb-4058-bb20-bc99dabaff31","Type":"ContainerDied","Data":"c30f87b0e6e5cd23e0d9ec3a789d10f56a2869b928021d324e962b976f2e57f6"} Feb 18 15:11:01 crc kubenswrapper[4896]: I0218 15:11:01.574773 4896 scope.go:117] "RemoveContainer" containerID="439e6446916aa08472620981e51299474c705bcd7ee0edd555ca02c9d2b791fe" Feb 18 15:11:01 crc kubenswrapper[4896]: I0218 15:11:01.574925 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dflvt" Feb 18 15:11:01 crc kubenswrapper[4896]: I0218 15:11:01.593707 4896 scope.go:117] "RemoveContainer" containerID="ff040b87961097f6a78477f8123d24f1fbf6845392f8d7f43a848144c0235a8f" Feb 18 15:11:01 crc kubenswrapper[4896]: I0218 15:11:01.600225 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dflvt"] Feb 18 15:11:01 crc kubenswrapper[4896]: I0218 15:11:01.606451 4896 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dflvt"] Feb 18 15:11:01 crc kubenswrapper[4896]: I0218 15:11:01.630219 4896 scope.go:117] "RemoveContainer" containerID="00a1069b88972b7541065703962ad5e6badbb9e6bc86b8ed9ac500a4b27cc2bf" Feb 18 15:11:01 crc kubenswrapper[4896]: I0218 15:11:01.647581 4896 scope.go:117] "RemoveContainer" containerID="439e6446916aa08472620981e51299474c705bcd7ee0edd555ca02c9d2b791fe" Feb 18 15:11:01 crc kubenswrapper[4896]: E0218 15:11:01.647918 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"439e6446916aa08472620981e51299474c705bcd7ee0edd555ca02c9d2b791fe\": container with ID starting with 439e6446916aa08472620981e51299474c705bcd7ee0edd555ca02c9d2b791fe not found: ID does not exist" containerID="439e6446916aa08472620981e51299474c705bcd7ee0edd555ca02c9d2b791fe" Feb 18 15:11:01 crc kubenswrapper[4896]: I0218 15:11:01.647948 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"439e6446916aa08472620981e51299474c705bcd7ee0edd555ca02c9d2b791fe"} err="failed to get container status \"439e6446916aa08472620981e51299474c705bcd7ee0edd555ca02c9d2b791fe\": rpc error: code = NotFound desc = could not find container \"439e6446916aa08472620981e51299474c705bcd7ee0edd555ca02c9d2b791fe\": container with ID starting with 439e6446916aa08472620981e51299474c705bcd7ee0edd555ca02c9d2b791fe not found: ID does not exist" Feb 18 15:11:01 crc kubenswrapper[4896]: I0218 15:11:01.647975 4896 scope.go:117] "RemoveContainer" containerID="ff040b87961097f6a78477f8123d24f1fbf6845392f8d7f43a848144c0235a8f" Feb 18 15:11:01 crc kubenswrapper[4896]: E0218 15:11:01.648281 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff040b87961097f6a78477f8123d24f1fbf6845392f8d7f43a848144c0235a8f\": container with ID starting with ff040b87961097f6a78477f8123d24f1fbf6845392f8d7f43a848144c0235a8f not found: ID does not exist" containerID="ff040b87961097f6a78477f8123d24f1fbf6845392f8d7f43a848144c0235a8f" Feb 18 15:11:01 crc kubenswrapper[4896]: I0218 15:11:01.648327 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff040b87961097f6a78477f8123d24f1fbf6845392f8d7f43a848144c0235a8f"} err="failed to get container status \"ff040b87961097f6a78477f8123d24f1fbf6845392f8d7f43a848144c0235a8f\": rpc error: code = NotFound desc = could not find container \"ff040b87961097f6a78477f8123d24f1fbf6845392f8d7f43a848144c0235a8f\": container with ID starting with ff040b87961097f6a78477f8123d24f1fbf6845392f8d7f43a848144c0235a8f not found: ID does not exist" Feb 18 15:11:01 crc kubenswrapper[4896]: I0218 15:11:01.648354 4896 scope.go:117] "RemoveContainer" containerID="00a1069b88972b7541065703962ad5e6badbb9e6bc86b8ed9ac500a4b27cc2bf" Feb 18 15:11:01 crc kubenswrapper[4896]: E0218 15:11:01.648621 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00a1069b88972b7541065703962ad5e6badbb9e6bc86b8ed9ac500a4b27cc2bf\": container with ID starting with 00a1069b88972b7541065703962ad5e6badbb9e6bc86b8ed9ac500a4b27cc2bf not found: ID does not exist" containerID="00a1069b88972b7541065703962ad5e6badbb9e6bc86b8ed9ac500a4b27cc2bf" Feb 18 15:11:01 crc kubenswrapper[4896]: I0218 15:11:01.648648 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00a1069b88972b7541065703962ad5e6badbb9e6bc86b8ed9ac500a4b27cc2bf"} err="failed to get container status \"00a1069b88972b7541065703962ad5e6badbb9e6bc86b8ed9ac500a4b27cc2bf\": rpc error: code = NotFound desc = could not find container \"00a1069b88972b7541065703962ad5e6badbb9e6bc86b8ed9ac500a4b27cc2bf\": container with ID starting with 00a1069b88972b7541065703962ad5e6badbb9e6bc86b8ed9ac500a4b27cc2bf not found: ID does not exist" Feb 18 15:11:01 crc kubenswrapper[4896]: I0218 15:11:01.935541 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f328bf3-bacb-4058-bb20-bc99dabaff31" path="/var/lib/kubelet/pods/9f328bf3-bacb-4058-bb20-bc99dabaff31/volumes" Feb 18 15:11:04 crc kubenswrapper[4896]: I0218 15:11:04.080819 4896 patch_prober.go:28] interesting pod/machine-config-daemon-24rq5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:11:04 crc kubenswrapper[4896]: I0218 15:11:04.081133 4896 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:11:04 crc kubenswrapper[4896]: I0218 15:11:04.081181 4896 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" Feb 18 15:11:04 crc kubenswrapper[4896]: I0218 15:11:04.081891 4896 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"763c989180a0be50e9c5a0c490b89d9b8795e413a592c48ba235f513fd0066bc"} pod="openshift-machine-config-operator/machine-config-daemon-24rq5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 18 15:11:04 crc kubenswrapper[4896]: I0218 15:11:04.081961 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" containerID="cri-o://763c989180a0be50e9c5a0c490b89d9b8795e413a592c48ba235f513fd0066bc" gracePeriod=600 Feb 18 15:11:04 crc kubenswrapper[4896]: I0218 15:11:04.384910 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="test-operator-kuttl-tests/rabbitmq-server-0" Feb 18 15:11:04 crc kubenswrapper[4896]: I0218 15:11:04.593442 4896 generic.go:334] "Generic (PLEG): container finished" podID="8503734a-8111-4fab-beda-1f0f3b59615b" containerID="763c989180a0be50e9c5a0c490b89d9b8795e413a592c48ba235f513fd0066bc" exitCode=0 Feb 18 15:11:04 crc kubenswrapper[4896]: I0218 15:11:04.593527 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" event={"ID":"8503734a-8111-4fab-beda-1f0f3b59615b","Type":"ContainerDied","Data":"763c989180a0be50e9c5a0c490b89d9b8795e413a592c48ba235f513fd0066bc"} Feb 18 15:11:04 crc kubenswrapper[4896]: I0218 15:11:04.593728 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" event={"ID":"8503734a-8111-4fab-beda-1f0f3b59615b","Type":"ContainerStarted","Data":"1c40add7af533d9a1676f813ee5909cf520a8eec67c80f8dbfffb153fcc0d39e"} Feb 18 15:11:04 crc kubenswrapper[4896]: I0218 15:11:04.593752 4896 scope.go:117] "RemoveContainer" containerID="91d6b5d30beb14a038ea309b143bbc7367c3406254b6275cdfed44c037d67a19" Feb 18 15:11:04 crc kubenswrapper[4896]: I0218 15:11:04.673557 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-77bdc966f6-nnk54" Feb 18 15:11:08 crc kubenswrapper[4896]: I0218 15:11:08.398329 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["test-operator-kuttl-tests/keystone-db-create-h8qzx"] Feb 18 15:11:08 crc kubenswrapper[4896]: E0218 15:11:08.399193 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f328bf3-bacb-4058-bb20-bc99dabaff31" containerName="extract-content" Feb 18 15:11:08 crc kubenswrapper[4896]: I0218 15:11:08.399211 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f328bf3-bacb-4058-bb20-bc99dabaff31" containerName="extract-content" Feb 18 15:11:08 crc kubenswrapper[4896]: E0218 15:11:08.399221 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f328bf3-bacb-4058-bb20-bc99dabaff31" containerName="registry-server" Feb 18 15:11:08 crc kubenswrapper[4896]: I0218 15:11:08.399246 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f328bf3-bacb-4058-bb20-bc99dabaff31" containerName="registry-server" Feb 18 15:11:08 crc kubenswrapper[4896]: E0218 15:11:08.399259 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f328bf3-bacb-4058-bb20-bc99dabaff31" containerName="extract-utilities" Feb 18 15:11:08 crc kubenswrapper[4896]: I0218 15:11:08.399269 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f328bf3-bacb-4058-bb20-bc99dabaff31" containerName="extract-utilities" Feb 18 15:11:08 crc kubenswrapper[4896]: I0218 15:11:08.399423 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f328bf3-bacb-4058-bb20-bc99dabaff31" containerName="registry-server" Feb 18 15:11:08 crc kubenswrapper[4896]: I0218 15:11:08.400460 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="test-operator-kuttl-tests/keystone-db-create-h8qzx" Feb 18 15:11:08 crc kubenswrapper[4896]: I0218 15:11:08.418987 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["test-operator-kuttl-tests/keystone-20ea-account-create-update-9qws6"] Feb 18 15:11:08 crc kubenswrapper[4896]: I0218 15:11:08.420146 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="test-operator-kuttl-tests/keystone-20ea-account-create-update-9qws6" Feb 18 15:11:08 crc kubenswrapper[4896]: I0218 15:11:08.428433 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["test-operator-kuttl-tests/keystone-db-create-h8qzx"] Feb 18 15:11:08 crc kubenswrapper[4896]: I0218 15:11:08.429767 4896 reflector.go:368] Caches populated for *v1.Secret from object-"test-operator-kuttl-tests"/"keystone-db-secret" Feb 18 15:11:08 crc kubenswrapper[4896]: I0218 15:11:08.433876 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["test-operator-kuttl-tests/keystone-20ea-account-create-update-9qws6"] Feb 18 15:11:08 crc kubenswrapper[4896]: I0218 15:11:08.515491 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65ed11b1-ec28-4b2a-b755-6d324c33426c-operator-scripts\") pod \"keystone-db-create-h8qzx\" (UID: \"65ed11b1-ec28-4b2a-b755-6d324c33426c\") " pod="test-operator-kuttl-tests/keystone-db-create-h8qzx" Feb 18 15:11:08 crc kubenswrapper[4896]: I0218 15:11:08.515543 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnfxk\" (UniqueName: \"kubernetes.io/projected/65ed11b1-ec28-4b2a-b755-6d324c33426c-kube-api-access-lnfxk\") pod \"keystone-db-create-h8qzx\" (UID: \"65ed11b1-ec28-4b2a-b755-6d324c33426c\") " pod="test-operator-kuttl-tests/keystone-db-create-h8qzx" Feb 18 15:11:08 crc kubenswrapper[4896]: I0218 15:11:08.616813 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65ed11b1-ec28-4b2a-b755-6d324c33426c-operator-scripts\") pod \"keystone-db-create-h8qzx\" (UID: \"65ed11b1-ec28-4b2a-b755-6d324c33426c\") " pod="test-operator-kuttl-tests/keystone-db-create-h8qzx" Feb 18 15:11:08 crc kubenswrapper[4896]: I0218 15:11:08.616854 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnfxk\" (UniqueName: \"kubernetes.io/projected/65ed11b1-ec28-4b2a-b755-6d324c33426c-kube-api-access-lnfxk\") pod \"keystone-db-create-h8qzx\" (UID: \"65ed11b1-ec28-4b2a-b755-6d324c33426c\") " pod="test-operator-kuttl-tests/keystone-db-create-h8qzx" Feb 18 15:11:08 crc kubenswrapper[4896]: I0218 15:11:08.616901 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc590030-d392-4cd4-96f5-d28039c0faa6-operator-scripts\") pod \"keystone-20ea-account-create-update-9qws6\" (UID: \"dc590030-d392-4cd4-96f5-d28039c0faa6\") " pod="test-operator-kuttl-tests/keystone-20ea-account-create-update-9qws6" Feb 18 15:11:08 crc kubenswrapper[4896]: I0218 15:11:08.616945 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdq2r\" (UniqueName: \"kubernetes.io/projected/dc590030-d392-4cd4-96f5-d28039c0faa6-kube-api-access-hdq2r\") pod \"keystone-20ea-account-create-update-9qws6\" (UID: \"dc590030-d392-4cd4-96f5-d28039c0faa6\") " pod="test-operator-kuttl-tests/keystone-20ea-account-create-update-9qws6" Feb 18 15:11:08 crc kubenswrapper[4896]: I0218 15:11:08.617840 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65ed11b1-ec28-4b2a-b755-6d324c33426c-operator-scripts\") pod \"keystone-db-create-h8qzx\" (UID: \"65ed11b1-ec28-4b2a-b755-6d324c33426c\") " pod="test-operator-kuttl-tests/keystone-db-create-h8qzx" Feb 18 15:11:08 crc kubenswrapper[4896]: I0218 15:11:08.648476 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnfxk\" (UniqueName: \"kubernetes.io/projected/65ed11b1-ec28-4b2a-b755-6d324c33426c-kube-api-access-lnfxk\") pod \"keystone-db-create-h8qzx\" (UID: \"65ed11b1-ec28-4b2a-b755-6d324c33426c\") " pod="test-operator-kuttl-tests/keystone-db-create-h8qzx" Feb 18 15:11:08 crc kubenswrapper[4896]: I0218 15:11:08.718106 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc590030-d392-4cd4-96f5-d28039c0faa6-operator-scripts\") pod \"keystone-20ea-account-create-update-9qws6\" (UID: \"dc590030-d392-4cd4-96f5-d28039c0faa6\") " pod="test-operator-kuttl-tests/keystone-20ea-account-create-update-9qws6" Feb 18 15:11:08 crc kubenswrapper[4896]: I0218 15:11:08.718199 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdq2r\" (UniqueName: \"kubernetes.io/projected/dc590030-d392-4cd4-96f5-d28039c0faa6-kube-api-access-hdq2r\") pod \"keystone-20ea-account-create-update-9qws6\" (UID: \"dc590030-d392-4cd4-96f5-d28039c0faa6\") " pod="test-operator-kuttl-tests/keystone-20ea-account-create-update-9qws6" Feb 18 15:11:08 crc kubenswrapper[4896]: I0218 15:11:08.718879 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc590030-d392-4cd4-96f5-d28039c0faa6-operator-scripts\") pod \"keystone-20ea-account-create-update-9qws6\" (UID: \"dc590030-d392-4cd4-96f5-d28039c0faa6\") " pod="test-operator-kuttl-tests/keystone-20ea-account-create-update-9qws6" Feb 18 15:11:08 crc kubenswrapper[4896]: I0218 15:11:08.735411 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="test-operator-kuttl-tests/keystone-db-create-h8qzx" Feb 18 15:11:08 crc kubenswrapper[4896]: I0218 15:11:08.735556 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdq2r\" (UniqueName: \"kubernetes.io/projected/dc590030-d392-4cd4-96f5-d28039c0faa6-kube-api-access-hdq2r\") pod \"keystone-20ea-account-create-update-9qws6\" (UID: \"dc590030-d392-4cd4-96f5-d28039c0faa6\") " pod="test-operator-kuttl-tests/keystone-20ea-account-create-update-9qws6" Feb 18 15:11:08 crc kubenswrapper[4896]: I0218 15:11:08.744703 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="test-operator-kuttl-tests/keystone-20ea-account-create-update-9qws6" Feb 18 15:11:09 crc kubenswrapper[4896]: I0218 15:11:09.144018 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["test-operator-kuttl-tests/keystone-db-create-h8qzx"] Feb 18 15:11:09 crc kubenswrapper[4896]: I0218 15:11:09.206551 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["test-operator-kuttl-tests/keystone-20ea-account-create-update-9qws6"] Feb 18 15:11:09 crc kubenswrapper[4896]: W0218 15:11:09.213485 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc590030_d392_4cd4_96f5_d28039c0faa6.slice/crio-9ee5011012db8e74011fc749d07e58778f030b446cdc31887f057c0010b8b710 WatchSource:0}: Error finding container 9ee5011012db8e74011fc749d07e58778f030b446cdc31887f057c0010b8b710: Status 404 returned error can't find the container with id 9ee5011012db8e74011fc749d07e58778f030b446cdc31887f057c0010b8b710 Feb 18 15:11:09 crc kubenswrapper[4896]: I0218 15:11:09.625412 4896 generic.go:334] "Generic (PLEG): container finished" podID="dc590030-d392-4cd4-96f5-d28039c0faa6" containerID="a3f0a1d2ff900601ef552bacc5623df47e20fe7663f70facd77aa511ce523a4f" exitCode=0 Feb 18 15:11:09 crc kubenswrapper[4896]: I0218 15:11:09.625609 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/keystone-20ea-account-create-update-9qws6" event={"ID":"dc590030-d392-4cd4-96f5-d28039c0faa6","Type":"ContainerDied","Data":"a3f0a1d2ff900601ef552bacc5623df47e20fe7663f70facd77aa511ce523a4f"} Feb 18 15:11:09 crc kubenswrapper[4896]: I0218 15:11:09.626765 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/keystone-20ea-account-create-update-9qws6" event={"ID":"dc590030-d392-4cd4-96f5-d28039c0faa6","Type":"ContainerStarted","Data":"9ee5011012db8e74011fc749d07e58778f030b446cdc31887f057c0010b8b710"} Feb 18 15:11:09 crc kubenswrapper[4896]: I0218 15:11:09.629834 4896 generic.go:334] "Generic (PLEG): container finished" podID="65ed11b1-ec28-4b2a-b755-6d324c33426c" containerID="27c80acd51b5f76959220e73178654655b4ae5f00a7245e3859fbac6aa5c83c8" exitCode=0 Feb 18 15:11:09 crc kubenswrapper[4896]: I0218 15:11:09.629885 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/keystone-db-create-h8qzx" event={"ID":"65ed11b1-ec28-4b2a-b755-6d324c33426c","Type":"ContainerDied","Data":"27c80acd51b5f76959220e73178654655b4ae5f00a7245e3859fbac6aa5c83c8"} Feb 18 15:11:09 crc kubenswrapper[4896]: I0218 15:11:09.629916 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/keystone-db-create-h8qzx" event={"ID":"65ed11b1-ec28-4b2a-b755-6d324c33426c","Type":"ContainerStarted","Data":"5b5db6aa188a0afe163c69e09b35476f8027fbe62fd5cd38cf6a2c67340349b7"} Feb 18 15:11:10 crc kubenswrapper[4896]: I0218 15:11:10.978431 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="test-operator-kuttl-tests/keystone-20ea-account-create-update-9qws6" Feb 18 15:11:10 crc kubenswrapper[4896]: I0218 15:11:10.985391 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="test-operator-kuttl-tests/keystone-db-create-h8qzx" Feb 18 15:11:11 crc kubenswrapper[4896]: I0218 15:11:11.147853 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnfxk\" (UniqueName: \"kubernetes.io/projected/65ed11b1-ec28-4b2a-b755-6d324c33426c-kube-api-access-lnfxk\") pod \"65ed11b1-ec28-4b2a-b755-6d324c33426c\" (UID: \"65ed11b1-ec28-4b2a-b755-6d324c33426c\") " Feb 18 15:11:11 crc kubenswrapper[4896]: I0218 15:11:11.148201 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65ed11b1-ec28-4b2a-b755-6d324c33426c-operator-scripts\") pod \"65ed11b1-ec28-4b2a-b755-6d324c33426c\" (UID: \"65ed11b1-ec28-4b2a-b755-6d324c33426c\") " Feb 18 15:11:11 crc kubenswrapper[4896]: I0218 15:11:11.148265 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdq2r\" (UniqueName: \"kubernetes.io/projected/dc590030-d392-4cd4-96f5-d28039c0faa6-kube-api-access-hdq2r\") pod \"dc590030-d392-4cd4-96f5-d28039c0faa6\" (UID: \"dc590030-d392-4cd4-96f5-d28039c0faa6\") " Feb 18 15:11:11 crc kubenswrapper[4896]: I0218 15:11:11.148381 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc590030-d392-4cd4-96f5-d28039c0faa6-operator-scripts\") pod \"dc590030-d392-4cd4-96f5-d28039c0faa6\" (UID: \"dc590030-d392-4cd4-96f5-d28039c0faa6\") " Feb 18 15:11:11 crc kubenswrapper[4896]: I0218 15:11:11.148957 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65ed11b1-ec28-4b2a-b755-6d324c33426c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "65ed11b1-ec28-4b2a-b755-6d324c33426c" (UID: "65ed11b1-ec28-4b2a-b755-6d324c33426c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:11:11 crc kubenswrapper[4896]: I0218 15:11:11.149180 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc590030-d392-4cd4-96f5-d28039c0faa6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dc590030-d392-4cd4-96f5-d28039c0faa6" (UID: "dc590030-d392-4cd4-96f5-d28039c0faa6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:11:11 crc kubenswrapper[4896]: I0218 15:11:11.154955 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65ed11b1-ec28-4b2a-b755-6d324c33426c-kube-api-access-lnfxk" (OuterVolumeSpecName: "kube-api-access-lnfxk") pod "65ed11b1-ec28-4b2a-b755-6d324c33426c" (UID: "65ed11b1-ec28-4b2a-b755-6d324c33426c"). InnerVolumeSpecName "kube-api-access-lnfxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:11:11 crc kubenswrapper[4896]: I0218 15:11:11.157374 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc590030-d392-4cd4-96f5-d28039c0faa6-kube-api-access-hdq2r" (OuterVolumeSpecName: "kube-api-access-hdq2r") pod "dc590030-d392-4cd4-96f5-d28039c0faa6" (UID: "dc590030-d392-4cd4-96f5-d28039c0faa6"). InnerVolumeSpecName "kube-api-access-hdq2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:11:11 crc kubenswrapper[4896]: I0218 15:11:11.250310 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnfxk\" (UniqueName: \"kubernetes.io/projected/65ed11b1-ec28-4b2a-b755-6d324c33426c-kube-api-access-lnfxk\") on node \"crc\" DevicePath \"\"" Feb 18 15:11:11 crc kubenswrapper[4896]: I0218 15:11:11.250349 4896 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/65ed11b1-ec28-4b2a-b755-6d324c33426c-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 15:11:11 crc kubenswrapper[4896]: I0218 15:11:11.250359 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdq2r\" (UniqueName: \"kubernetes.io/projected/dc590030-d392-4cd4-96f5-d28039c0faa6-kube-api-access-hdq2r\") on node \"crc\" DevicePath \"\"" Feb 18 15:11:11 crc kubenswrapper[4896]: I0218 15:11:11.250367 4896 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc590030-d392-4cd4-96f5-d28039c0faa6-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 15:11:11 crc kubenswrapper[4896]: I0218 15:11:11.645923 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="test-operator-kuttl-tests/keystone-20ea-account-create-update-9qws6" Feb 18 15:11:11 crc kubenswrapper[4896]: I0218 15:11:11.645922 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/keystone-20ea-account-create-update-9qws6" event={"ID":"dc590030-d392-4cd4-96f5-d28039c0faa6","Type":"ContainerDied","Data":"9ee5011012db8e74011fc749d07e58778f030b446cdc31887f057c0010b8b710"} Feb 18 15:11:11 crc kubenswrapper[4896]: I0218 15:11:11.646289 4896 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ee5011012db8e74011fc749d07e58778f030b446cdc31887f057c0010b8b710" Feb 18 15:11:11 crc kubenswrapper[4896]: I0218 15:11:11.647290 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="test-operator-kuttl-tests/keystone-db-create-h8qzx" Feb 18 15:11:11 crc kubenswrapper[4896]: I0218 15:11:11.647336 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/keystone-db-create-h8qzx" event={"ID":"65ed11b1-ec28-4b2a-b755-6d324c33426c","Type":"ContainerDied","Data":"5b5db6aa188a0afe163c69e09b35476f8027fbe62fd5cd38cf6a2c67340349b7"} Feb 18 15:11:11 crc kubenswrapper[4896]: I0218 15:11:11.647369 4896 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b5db6aa188a0afe163c69e09b35476f8027fbe62fd5cd38cf6a2c67340349b7" Feb 18 15:11:12 crc kubenswrapper[4896]: I0218 15:11:12.280086 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-operator-index-ngrc4"] Feb 18 15:11:12 crc kubenswrapper[4896]: E0218 15:11:12.280380 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc590030-d392-4cd4-96f5-d28039c0faa6" containerName="mariadb-account-create-update" Feb 18 15:11:12 crc kubenswrapper[4896]: I0218 15:11:12.280395 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc590030-d392-4cd4-96f5-d28039c0faa6" containerName="mariadb-account-create-update" Feb 18 15:11:12 crc kubenswrapper[4896]: E0218 15:11:12.280407 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65ed11b1-ec28-4b2a-b755-6d324c33426c" containerName="mariadb-database-create" Feb 18 15:11:12 crc kubenswrapper[4896]: I0218 15:11:12.280418 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="65ed11b1-ec28-4b2a-b755-6d324c33426c" containerName="mariadb-database-create" Feb 18 15:11:12 crc kubenswrapper[4896]: I0218 15:11:12.280580 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc590030-d392-4cd4-96f5-d28039c0faa6" containerName="mariadb-account-create-update" Feb 18 15:11:12 crc kubenswrapper[4896]: I0218 15:11:12.280598 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="65ed11b1-ec28-4b2a-b755-6d324c33426c" containerName="mariadb-database-create" Feb 18 15:11:12 crc kubenswrapper[4896]: I0218 15:11:12.281132 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-operator-index-ngrc4" Feb 18 15:11:12 crc kubenswrapper[4896]: I0218 15:11:12.283319 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-operator-index-dockercfg-4tbms" Feb 18 15:11:12 crc kubenswrapper[4896]: I0218 15:11:12.287906 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-operator-index-ngrc4"] Feb 18 15:11:12 crc kubenswrapper[4896]: I0218 15:11:12.365489 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bp68\" (UniqueName: \"kubernetes.io/projected/5962f084-f27d-47b3-a9c1-b1389078c88b-kube-api-access-4bp68\") pod \"test-operator-operator-index-ngrc4\" (UID: \"5962f084-f27d-47b3-a9c1-b1389078c88b\") " pod="openstack-operators/test-operator-operator-index-ngrc4" Feb 18 15:11:12 crc kubenswrapper[4896]: I0218 15:11:12.467129 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bp68\" (UniqueName: \"kubernetes.io/projected/5962f084-f27d-47b3-a9c1-b1389078c88b-kube-api-access-4bp68\") pod \"test-operator-operator-index-ngrc4\" (UID: \"5962f084-f27d-47b3-a9c1-b1389078c88b\") " pod="openstack-operators/test-operator-operator-index-ngrc4" Feb 18 15:11:12 crc kubenswrapper[4896]: I0218 15:11:12.487245 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bp68\" (UniqueName: \"kubernetes.io/projected/5962f084-f27d-47b3-a9c1-b1389078c88b-kube-api-access-4bp68\") pod \"test-operator-operator-index-ngrc4\" (UID: \"5962f084-f27d-47b3-a9c1-b1389078c88b\") " pod="openstack-operators/test-operator-operator-index-ngrc4" Feb 18 15:11:12 crc kubenswrapper[4896]: I0218 15:11:12.605637 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-operator-index-ngrc4" Feb 18 15:11:13 crc kubenswrapper[4896]: I0218 15:11:13.041789 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-operator-index-ngrc4"] Feb 18 15:11:13 crc kubenswrapper[4896]: I0218 15:11:13.668032 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-operator-index-ngrc4" event={"ID":"5962f084-f27d-47b3-a9c1-b1389078c88b","Type":"ContainerStarted","Data":"1c2e4dafff039c16d6a9b68c90dbc6ebe33c02cea34ce4942b5b7ffaadf586ec"} Feb 18 15:11:13 crc kubenswrapper[4896]: I0218 15:11:13.951457 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["test-operator-kuttl-tests/keystone-db-sync-nf7cx"] Feb 18 15:11:13 crc kubenswrapper[4896]: I0218 15:11:13.952419 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="test-operator-kuttl-tests/keystone-db-sync-nf7cx" Feb 18 15:11:13 crc kubenswrapper[4896]: I0218 15:11:13.954382 4896 reflector.go:368] Caches populated for *v1.Secret from object-"test-operator-kuttl-tests"/"keystone" Feb 18 15:11:13 crc kubenswrapper[4896]: I0218 15:11:13.955187 4896 reflector.go:368] Caches populated for *v1.Secret from object-"test-operator-kuttl-tests"/"keystone-config-data" Feb 18 15:11:13 crc kubenswrapper[4896]: I0218 15:11:13.955354 4896 reflector.go:368] Caches populated for *v1.Secret from object-"test-operator-kuttl-tests"/"keystone-keystone-dockercfg-vdf5c" Feb 18 15:11:13 crc kubenswrapper[4896]: I0218 15:11:13.955364 4896 reflector.go:368] Caches populated for *v1.Secret from object-"test-operator-kuttl-tests"/"keystone-scripts" Feb 18 15:11:13 crc kubenswrapper[4896]: I0218 15:11:13.956974 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["test-operator-kuttl-tests/keystone-db-sync-nf7cx"] Feb 18 15:11:13 crc kubenswrapper[4896]: I0218 15:11:13.989065 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74496765-f69d-4a41-add9-5afe29af7e20-config-data\") pod \"keystone-db-sync-nf7cx\" (UID: \"74496765-f69d-4a41-add9-5afe29af7e20\") " pod="test-operator-kuttl-tests/keystone-db-sync-nf7cx" Feb 18 15:11:13 crc kubenswrapper[4896]: I0218 15:11:13.989203 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mch8z\" (UniqueName: \"kubernetes.io/projected/74496765-f69d-4a41-add9-5afe29af7e20-kube-api-access-mch8z\") pod \"keystone-db-sync-nf7cx\" (UID: \"74496765-f69d-4a41-add9-5afe29af7e20\") " pod="test-operator-kuttl-tests/keystone-db-sync-nf7cx" Feb 18 15:11:14 crc kubenswrapper[4896]: I0218 15:11:14.090204 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mch8z\" (UniqueName: \"kubernetes.io/projected/74496765-f69d-4a41-add9-5afe29af7e20-kube-api-access-mch8z\") pod \"keystone-db-sync-nf7cx\" (UID: \"74496765-f69d-4a41-add9-5afe29af7e20\") " pod="test-operator-kuttl-tests/keystone-db-sync-nf7cx" Feb 18 15:11:14 crc kubenswrapper[4896]: I0218 15:11:14.090304 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74496765-f69d-4a41-add9-5afe29af7e20-config-data\") pod \"keystone-db-sync-nf7cx\" (UID: \"74496765-f69d-4a41-add9-5afe29af7e20\") " pod="test-operator-kuttl-tests/keystone-db-sync-nf7cx" Feb 18 15:11:14 crc kubenswrapper[4896]: I0218 15:11:14.098443 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74496765-f69d-4a41-add9-5afe29af7e20-config-data\") pod \"keystone-db-sync-nf7cx\" (UID: \"74496765-f69d-4a41-add9-5afe29af7e20\") " pod="test-operator-kuttl-tests/keystone-db-sync-nf7cx" Feb 18 15:11:14 crc kubenswrapper[4896]: I0218 15:11:14.106574 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mch8z\" (UniqueName: \"kubernetes.io/projected/74496765-f69d-4a41-add9-5afe29af7e20-kube-api-access-mch8z\") pod \"keystone-db-sync-nf7cx\" (UID: \"74496765-f69d-4a41-add9-5afe29af7e20\") " pod="test-operator-kuttl-tests/keystone-db-sync-nf7cx" Feb 18 15:11:14 crc kubenswrapper[4896]: I0218 15:11:14.279431 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="test-operator-kuttl-tests/keystone-db-sync-nf7cx" Feb 18 15:11:15 crc kubenswrapper[4896]: I0218 15:11:15.686654 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-operator-index-ngrc4" event={"ID":"5962f084-f27d-47b3-a9c1-b1389078c88b","Type":"ContainerStarted","Data":"0e843a040ebf9a041f44e358fb8ab2cac9dfd572f78f208c46849c4d2cfff82b"} Feb 18 15:11:15 crc kubenswrapper[4896]: I0218 15:11:15.703357 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-operator-index-ngrc4" podStartSLOduration=1.323797178 podStartE2EDuration="3.703338567s" podCreationTimestamp="2026-02-18 15:11:12 +0000 UTC" firstStartedPulling="2026-02-18 15:11:13.048009423 +0000 UTC m=+971.977343636" lastFinishedPulling="2026-02-18 15:11:15.427550812 +0000 UTC m=+974.356885025" observedRunningTime="2026-02-18 15:11:15.701118409 +0000 UTC m=+974.630452622" watchObservedRunningTime="2026-02-18 15:11:15.703338567 +0000 UTC m=+974.632672780" Feb 18 15:11:15 crc kubenswrapper[4896]: I0218 15:11:15.726572 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["test-operator-kuttl-tests/keystone-db-sync-nf7cx"] Feb 18 15:11:15 crc kubenswrapper[4896]: W0218 15:11:15.732499 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74496765_f69d_4a41_add9_5afe29af7e20.slice/crio-cbb5196704800370a22f056f518d2b39ceca9d4fde90638254971d34e584e667 WatchSource:0}: Error finding container cbb5196704800370a22f056f518d2b39ceca9d4fde90638254971d34e584e667: Status 404 returned error can't find the container with id cbb5196704800370a22f056f518d2b39ceca9d4fde90638254971d34e584e667 Feb 18 15:11:16 crc kubenswrapper[4896]: I0218 15:11:16.697733 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/keystone-db-sync-nf7cx" event={"ID":"74496765-f69d-4a41-add9-5afe29af7e20","Type":"ContainerStarted","Data":"cbb5196704800370a22f056f518d2b39ceca9d4fde90638254971d34e584e667"} Feb 18 15:11:22 crc kubenswrapper[4896]: I0218 15:11:22.606803 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/test-operator-operator-index-ngrc4" Feb 18 15:11:22 crc kubenswrapper[4896]: I0218 15:11:22.607396 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-operator-index-ngrc4" Feb 18 15:11:22 crc kubenswrapper[4896]: I0218 15:11:22.638485 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/test-operator-operator-index-ngrc4" Feb 18 15:11:22 crc kubenswrapper[4896]: I0218 15:11:22.763256 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-operator-index-ngrc4" Feb 18 15:11:23 crc kubenswrapper[4896]: I0218 15:11:23.758869 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/keystone-db-sync-nf7cx" event={"ID":"74496765-f69d-4a41-add9-5afe29af7e20","Type":"ContainerStarted","Data":"00f6a64d269d0990c6c301cf52110604d3bb223484458443ba64902717afae5b"} Feb 18 15:11:23 crc kubenswrapper[4896]: I0218 15:11:23.778009 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="test-operator-kuttl-tests/keystone-db-sync-nf7cx" podStartSLOduration=3.782027694 podStartE2EDuration="10.777989588s" podCreationTimestamp="2026-02-18 15:11:13 +0000 UTC" firstStartedPulling="2026-02-18 15:11:15.73664919 +0000 UTC m=+974.665983403" lastFinishedPulling="2026-02-18 15:11:22.732611084 +0000 UTC m=+981.661945297" observedRunningTime="2026-02-18 15:11:23.776824253 +0000 UTC m=+982.706158476" watchObservedRunningTime="2026-02-18 15:11:23.777989588 +0000 UTC m=+982.707323811" Feb 18 15:11:26 crc kubenswrapper[4896]: I0218 15:11:26.782686 4896 generic.go:334] "Generic (PLEG): container finished" podID="74496765-f69d-4a41-add9-5afe29af7e20" containerID="00f6a64d269d0990c6c301cf52110604d3bb223484458443ba64902717afae5b" exitCode=0 Feb 18 15:11:26 crc kubenswrapper[4896]: I0218 15:11:26.782733 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/keystone-db-sync-nf7cx" event={"ID":"74496765-f69d-4a41-add9-5afe29af7e20","Type":"ContainerDied","Data":"00f6a64d269d0990c6c301cf52110604d3bb223484458443ba64902717afae5b"} Feb 18 15:11:28 crc kubenswrapper[4896]: I0218 15:11:28.030206 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="test-operator-kuttl-tests/keystone-db-sync-nf7cx" Feb 18 15:11:28 crc kubenswrapper[4896]: I0218 15:11:28.199358 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mch8z\" (UniqueName: \"kubernetes.io/projected/74496765-f69d-4a41-add9-5afe29af7e20-kube-api-access-mch8z\") pod \"74496765-f69d-4a41-add9-5afe29af7e20\" (UID: \"74496765-f69d-4a41-add9-5afe29af7e20\") " Feb 18 15:11:28 crc kubenswrapper[4896]: I0218 15:11:28.199473 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74496765-f69d-4a41-add9-5afe29af7e20-config-data\") pod \"74496765-f69d-4a41-add9-5afe29af7e20\" (UID: \"74496765-f69d-4a41-add9-5afe29af7e20\") " Feb 18 15:11:28 crc kubenswrapper[4896]: I0218 15:11:28.210571 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74496765-f69d-4a41-add9-5afe29af7e20-kube-api-access-mch8z" (OuterVolumeSpecName: "kube-api-access-mch8z") pod "74496765-f69d-4a41-add9-5afe29af7e20" (UID: "74496765-f69d-4a41-add9-5afe29af7e20"). InnerVolumeSpecName "kube-api-access-mch8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:11:28 crc kubenswrapper[4896]: I0218 15:11:28.237821 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74496765-f69d-4a41-add9-5afe29af7e20-config-data" (OuterVolumeSpecName: "config-data") pod "74496765-f69d-4a41-add9-5afe29af7e20" (UID: "74496765-f69d-4a41-add9-5afe29af7e20"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:11:28 crc kubenswrapper[4896]: I0218 15:11:28.301429 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mch8z\" (UniqueName: \"kubernetes.io/projected/74496765-f69d-4a41-add9-5afe29af7e20-kube-api-access-mch8z\") on node \"crc\" DevicePath \"\"" Feb 18 15:11:28 crc kubenswrapper[4896]: I0218 15:11:28.301476 4896 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74496765-f69d-4a41-add9-5afe29af7e20-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 15:11:28 crc kubenswrapper[4896]: I0218 15:11:28.793542 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/keystone-db-sync-nf7cx" event={"ID":"74496765-f69d-4a41-add9-5afe29af7e20","Type":"ContainerDied","Data":"cbb5196704800370a22f056f518d2b39ceca9d4fde90638254971d34e584e667"} Feb 18 15:11:28 crc kubenswrapper[4896]: I0218 15:11:28.793590 4896 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cbb5196704800370a22f056f518d2b39ceca9d4fde90638254971d34e584e667" Feb 18 15:11:28 crc kubenswrapper[4896]: I0218 15:11:28.793590 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="test-operator-kuttl-tests/keystone-db-sync-nf7cx" Feb 18 15:11:29 crc kubenswrapper[4896]: I0218 15:11:29.003938 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["test-operator-kuttl-tests/keystone-bootstrap-bmd8m"] Feb 18 15:11:29 crc kubenswrapper[4896]: E0218 15:11:29.004213 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74496765-f69d-4a41-add9-5afe29af7e20" containerName="keystone-db-sync" Feb 18 15:11:29 crc kubenswrapper[4896]: I0218 15:11:29.004230 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="74496765-f69d-4a41-add9-5afe29af7e20" containerName="keystone-db-sync" Feb 18 15:11:29 crc kubenswrapper[4896]: I0218 15:11:29.004406 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="74496765-f69d-4a41-add9-5afe29af7e20" containerName="keystone-db-sync" Feb 18 15:11:29 crc kubenswrapper[4896]: I0218 15:11:29.004826 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="test-operator-kuttl-tests/keystone-bootstrap-bmd8m" Feb 18 15:11:29 crc kubenswrapper[4896]: I0218 15:11:29.008024 4896 reflector.go:368] Caches populated for *v1.Secret from object-"test-operator-kuttl-tests"/"keystone" Feb 18 15:11:29 crc kubenswrapper[4896]: I0218 15:11:29.008393 4896 reflector.go:368] Caches populated for *v1.Secret from object-"test-operator-kuttl-tests"/"osp-secret" Feb 18 15:11:29 crc kubenswrapper[4896]: I0218 15:11:29.008495 4896 reflector.go:368] Caches populated for *v1.Secret from object-"test-operator-kuttl-tests"/"keystone-scripts" Feb 18 15:11:29 crc kubenswrapper[4896]: I0218 15:11:29.008560 4896 reflector.go:368] Caches populated for *v1.Secret from object-"test-operator-kuttl-tests"/"keystone-config-data" Feb 18 15:11:29 crc kubenswrapper[4896]: I0218 15:11:29.008768 4896 reflector.go:368] Caches populated for *v1.Secret from object-"test-operator-kuttl-tests"/"keystone-keystone-dockercfg-vdf5c" Feb 18 15:11:29 crc kubenswrapper[4896]: I0218 15:11:29.019412 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["test-operator-kuttl-tests/keystone-bootstrap-bmd8m"] Feb 18 15:11:29 crc kubenswrapper[4896]: I0218 15:11:29.112877 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d53c04c-f334-4252-91e3-8feb31200ee7-config-data\") pod \"keystone-bootstrap-bmd8m\" (UID: \"2d53c04c-f334-4252-91e3-8feb31200ee7\") " pod="test-operator-kuttl-tests/keystone-bootstrap-bmd8m" Feb 18 15:11:29 crc kubenswrapper[4896]: I0218 15:11:29.113182 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d53c04c-f334-4252-91e3-8feb31200ee7-scripts\") pod \"keystone-bootstrap-bmd8m\" (UID: \"2d53c04c-f334-4252-91e3-8feb31200ee7\") " pod="test-operator-kuttl-tests/keystone-bootstrap-bmd8m" Feb 18 15:11:29 crc kubenswrapper[4896]: I0218 15:11:29.113248 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgw5s\" (UniqueName: \"kubernetes.io/projected/2d53c04c-f334-4252-91e3-8feb31200ee7-kube-api-access-vgw5s\") pod \"keystone-bootstrap-bmd8m\" (UID: \"2d53c04c-f334-4252-91e3-8feb31200ee7\") " pod="test-operator-kuttl-tests/keystone-bootstrap-bmd8m" Feb 18 15:11:29 crc kubenswrapper[4896]: I0218 15:11:29.113569 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2d53c04c-f334-4252-91e3-8feb31200ee7-fernet-keys\") pod \"keystone-bootstrap-bmd8m\" (UID: \"2d53c04c-f334-4252-91e3-8feb31200ee7\") " pod="test-operator-kuttl-tests/keystone-bootstrap-bmd8m" Feb 18 15:11:29 crc kubenswrapper[4896]: I0218 15:11:29.113738 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2d53c04c-f334-4252-91e3-8feb31200ee7-credential-keys\") pod \"keystone-bootstrap-bmd8m\" (UID: \"2d53c04c-f334-4252-91e3-8feb31200ee7\") " pod="test-operator-kuttl-tests/keystone-bootstrap-bmd8m" Feb 18 15:11:29 crc kubenswrapper[4896]: I0218 15:11:29.215583 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2d53c04c-f334-4252-91e3-8feb31200ee7-credential-keys\") pod \"keystone-bootstrap-bmd8m\" (UID: \"2d53c04c-f334-4252-91e3-8feb31200ee7\") " pod="test-operator-kuttl-tests/keystone-bootstrap-bmd8m" Feb 18 15:11:29 crc kubenswrapper[4896]: I0218 15:11:29.215656 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d53c04c-f334-4252-91e3-8feb31200ee7-config-data\") pod \"keystone-bootstrap-bmd8m\" (UID: \"2d53c04c-f334-4252-91e3-8feb31200ee7\") " pod="test-operator-kuttl-tests/keystone-bootstrap-bmd8m" Feb 18 15:11:29 crc kubenswrapper[4896]: I0218 15:11:29.215685 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d53c04c-f334-4252-91e3-8feb31200ee7-scripts\") pod \"keystone-bootstrap-bmd8m\" (UID: \"2d53c04c-f334-4252-91e3-8feb31200ee7\") " pod="test-operator-kuttl-tests/keystone-bootstrap-bmd8m" Feb 18 15:11:29 crc kubenswrapper[4896]: I0218 15:11:29.215733 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgw5s\" (UniqueName: \"kubernetes.io/projected/2d53c04c-f334-4252-91e3-8feb31200ee7-kube-api-access-vgw5s\") pod \"keystone-bootstrap-bmd8m\" (UID: \"2d53c04c-f334-4252-91e3-8feb31200ee7\") " pod="test-operator-kuttl-tests/keystone-bootstrap-bmd8m" Feb 18 15:11:29 crc kubenswrapper[4896]: I0218 15:11:29.215779 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2d53c04c-f334-4252-91e3-8feb31200ee7-fernet-keys\") pod \"keystone-bootstrap-bmd8m\" (UID: \"2d53c04c-f334-4252-91e3-8feb31200ee7\") " pod="test-operator-kuttl-tests/keystone-bootstrap-bmd8m" Feb 18 15:11:29 crc kubenswrapper[4896]: I0218 15:11:29.220678 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2d53c04c-f334-4252-91e3-8feb31200ee7-fernet-keys\") pod \"keystone-bootstrap-bmd8m\" (UID: \"2d53c04c-f334-4252-91e3-8feb31200ee7\") " pod="test-operator-kuttl-tests/keystone-bootstrap-bmd8m" Feb 18 15:11:29 crc kubenswrapper[4896]: I0218 15:11:29.220742 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d53c04c-f334-4252-91e3-8feb31200ee7-config-data\") pod \"keystone-bootstrap-bmd8m\" (UID: \"2d53c04c-f334-4252-91e3-8feb31200ee7\") " pod="test-operator-kuttl-tests/keystone-bootstrap-bmd8m" Feb 18 15:11:29 crc kubenswrapper[4896]: I0218 15:11:29.232162 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d53c04c-f334-4252-91e3-8feb31200ee7-scripts\") pod \"keystone-bootstrap-bmd8m\" (UID: \"2d53c04c-f334-4252-91e3-8feb31200ee7\") " pod="test-operator-kuttl-tests/keystone-bootstrap-bmd8m" Feb 18 15:11:29 crc kubenswrapper[4896]: I0218 15:11:29.233055 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2d53c04c-f334-4252-91e3-8feb31200ee7-credential-keys\") pod \"keystone-bootstrap-bmd8m\" (UID: \"2d53c04c-f334-4252-91e3-8feb31200ee7\") " pod="test-operator-kuttl-tests/keystone-bootstrap-bmd8m" Feb 18 15:11:29 crc kubenswrapper[4896]: I0218 15:11:29.239922 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgw5s\" (UniqueName: \"kubernetes.io/projected/2d53c04c-f334-4252-91e3-8feb31200ee7-kube-api-access-vgw5s\") pod \"keystone-bootstrap-bmd8m\" (UID: \"2d53c04c-f334-4252-91e3-8feb31200ee7\") " pod="test-operator-kuttl-tests/keystone-bootstrap-bmd8m" Feb 18 15:11:29 crc kubenswrapper[4896]: I0218 15:11:29.320413 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="test-operator-kuttl-tests/keystone-bootstrap-bmd8m" Feb 18 15:11:29 crc kubenswrapper[4896]: I0218 15:11:29.725553 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["test-operator-kuttl-tests/keystone-bootstrap-bmd8m"] Feb 18 15:11:29 crc kubenswrapper[4896]: W0218 15:11:29.733218 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d53c04c_f334_4252_91e3_8feb31200ee7.slice/crio-2f9254c09c57a8ec4f47cd4933edc968ff03c4533caf5ef12d2e58dcc65c5f30 WatchSource:0}: Error finding container 2f9254c09c57a8ec4f47cd4933edc968ff03c4533caf5ef12d2e58dcc65c5f30: Status 404 returned error can't find the container with id 2f9254c09c57a8ec4f47cd4933edc968ff03c4533caf5ef12d2e58dcc65c5f30 Feb 18 15:11:29 crc kubenswrapper[4896]: I0218 15:11:29.800611 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/keystone-bootstrap-bmd8m" event={"ID":"2d53c04c-f334-4252-91e3-8feb31200ee7","Type":"ContainerStarted","Data":"2f9254c09c57a8ec4f47cd4933edc968ff03c4533caf5ef12d2e58dcc65c5f30"} Feb 18 15:11:30 crc kubenswrapper[4896]: I0218 15:11:30.806708 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/keystone-bootstrap-bmd8m" event={"ID":"2d53c04c-f334-4252-91e3-8feb31200ee7","Type":"ContainerStarted","Data":"844eb879befbbb02991934eb32d29f0c96ab9a51a086248bf69a58c61f9b2d7a"} Feb 18 15:11:30 crc kubenswrapper[4896]: I0218 15:11:30.823784 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="test-operator-kuttl-tests/keystone-bootstrap-bmd8m" podStartSLOduration=2.823766643 podStartE2EDuration="2.823766643s" podCreationTimestamp="2026-02-18 15:11:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:11:30.819896219 +0000 UTC m=+989.749230432" watchObservedRunningTime="2026-02-18 15:11:30.823766643 +0000 UTC m=+989.753100856" Feb 18 15:11:33 crc kubenswrapper[4896]: I0218 15:11:33.824707 4896 generic.go:334] "Generic (PLEG): container finished" podID="2d53c04c-f334-4252-91e3-8feb31200ee7" containerID="844eb879befbbb02991934eb32d29f0c96ab9a51a086248bf69a58c61f9b2d7a" exitCode=0 Feb 18 15:11:33 crc kubenswrapper[4896]: I0218 15:11:33.824783 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/keystone-bootstrap-bmd8m" event={"ID":"2d53c04c-f334-4252-91e3-8feb31200ee7","Type":"ContainerDied","Data":"844eb879befbbb02991934eb32d29f0c96ab9a51a086248bf69a58c61f9b2d7a"} Feb 18 15:11:35 crc kubenswrapper[4896]: I0218 15:11:35.099405 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="test-operator-kuttl-tests/keystone-bootstrap-bmd8m" Feb 18 15:11:35 crc kubenswrapper[4896]: I0218 15:11:35.202897 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d53c04c-f334-4252-91e3-8feb31200ee7-scripts\") pod \"2d53c04c-f334-4252-91e3-8feb31200ee7\" (UID: \"2d53c04c-f334-4252-91e3-8feb31200ee7\") " Feb 18 15:11:35 crc kubenswrapper[4896]: I0218 15:11:35.202945 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d53c04c-f334-4252-91e3-8feb31200ee7-config-data\") pod \"2d53c04c-f334-4252-91e3-8feb31200ee7\" (UID: \"2d53c04c-f334-4252-91e3-8feb31200ee7\") " Feb 18 15:11:35 crc kubenswrapper[4896]: I0218 15:11:35.202965 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vgw5s\" (UniqueName: \"kubernetes.io/projected/2d53c04c-f334-4252-91e3-8feb31200ee7-kube-api-access-vgw5s\") pod \"2d53c04c-f334-4252-91e3-8feb31200ee7\" (UID: \"2d53c04c-f334-4252-91e3-8feb31200ee7\") " Feb 18 15:11:35 crc kubenswrapper[4896]: I0218 15:11:35.203016 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2d53c04c-f334-4252-91e3-8feb31200ee7-fernet-keys\") pod \"2d53c04c-f334-4252-91e3-8feb31200ee7\" (UID: \"2d53c04c-f334-4252-91e3-8feb31200ee7\") " Feb 18 15:11:35 crc kubenswrapper[4896]: I0218 15:11:35.203036 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2d53c04c-f334-4252-91e3-8feb31200ee7-credential-keys\") pod \"2d53c04c-f334-4252-91e3-8feb31200ee7\" (UID: \"2d53c04c-f334-4252-91e3-8feb31200ee7\") " Feb 18 15:11:35 crc kubenswrapper[4896]: I0218 15:11:35.209393 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d53c04c-f334-4252-91e3-8feb31200ee7-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "2d53c04c-f334-4252-91e3-8feb31200ee7" (UID: "2d53c04c-f334-4252-91e3-8feb31200ee7"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:11:35 crc kubenswrapper[4896]: I0218 15:11:35.209414 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d53c04c-f334-4252-91e3-8feb31200ee7-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "2d53c04c-f334-4252-91e3-8feb31200ee7" (UID: "2d53c04c-f334-4252-91e3-8feb31200ee7"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:11:35 crc kubenswrapper[4896]: I0218 15:11:35.209430 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d53c04c-f334-4252-91e3-8feb31200ee7-kube-api-access-vgw5s" (OuterVolumeSpecName: "kube-api-access-vgw5s") pod "2d53c04c-f334-4252-91e3-8feb31200ee7" (UID: "2d53c04c-f334-4252-91e3-8feb31200ee7"). InnerVolumeSpecName "kube-api-access-vgw5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:11:35 crc kubenswrapper[4896]: I0218 15:11:35.209441 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d53c04c-f334-4252-91e3-8feb31200ee7-scripts" (OuterVolumeSpecName: "scripts") pod "2d53c04c-f334-4252-91e3-8feb31200ee7" (UID: "2d53c04c-f334-4252-91e3-8feb31200ee7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:11:35 crc kubenswrapper[4896]: I0218 15:11:35.221697 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d53c04c-f334-4252-91e3-8feb31200ee7-config-data" (OuterVolumeSpecName: "config-data") pod "2d53c04c-f334-4252-91e3-8feb31200ee7" (UID: "2d53c04c-f334-4252-91e3-8feb31200ee7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:11:35 crc kubenswrapper[4896]: I0218 15:11:35.307611 4896 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d53c04c-f334-4252-91e3-8feb31200ee7-scripts\") on node \"crc\" DevicePath \"\"" Feb 18 15:11:35 crc kubenswrapper[4896]: I0218 15:11:35.307864 4896 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d53c04c-f334-4252-91e3-8feb31200ee7-config-data\") on node \"crc\" DevicePath \"\"" Feb 18 15:11:35 crc kubenswrapper[4896]: I0218 15:11:35.307882 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vgw5s\" (UniqueName: \"kubernetes.io/projected/2d53c04c-f334-4252-91e3-8feb31200ee7-kube-api-access-vgw5s\") on node \"crc\" DevicePath \"\"" Feb 18 15:11:35 crc kubenswrapper[4896]: I0218 15:11:35.307897 4896 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2d53c04c-f334-4252-91e3-8feb31200ee7-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 18 15:11:35 crc kubenswrapper[4896]: I0218 15:11:35.307907 4896 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2d53c04c-f334-4252-91e3-8feb31200ee7-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 18 15:11:35 crc kubenswrapper[4896]: I0218 15:11:35.837212 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/keystone-bootstrap-bmd8m" event={"ID":"2d53c04c-f334-4252-91e3-8feb31200ee7","Type":"ContainerDied","Data":"2f9254c09c57a8ec4f47cd4933edc968ff03c4533caf5ef12d2e58dcc65c5f30"} Feb 18 15:11:35 crc kubenswrapper[4896]: I0218 15:11:35.837572 4896 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f9254c09c57a8ec4f47cd4933edc968ff03c4533caf5ef12d2e58dcc65c5f30" Feb 18 15:11:35 crc kubenswrapper[4896]: I0218 15:11:35.837378 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="test-operator-kuttl-tests/keystone-bootstrap-bmd8m" Feb 18 15:11:35 crc kubenswrapper[4896]: I0218 15:11:35.909803 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["test-operator-kuttl-tests/keystone-7454c655dc-bjql4"] Feb 18 15:11:35 crc kubenswrapper[4896]: E0218 15:11:35.910049 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d53c04c-f334-4252-91e3-8feb31200ee7" containerName="keystone-bootstrap" Feb 18 15:11:35 crc kubenswrapper[4896]: I0218 15:11:35.910066 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d53c04c-f334-4252-91e3-8feb31200ee7" containerName="keystone-bootstrap" Feb 18 15:11:35 crc kubenswrapper[4896]: I0218 15:11:35.910180 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d53c04c-f334-4252-91e3-8feb31200ee7" containerName="keystone-bootstrap" Feb 18 15:11:35 crc kubenswrapper[4896]: I0218 15:11:35.910796 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="test-operator-kuttl-tests/keystone-7454c655dc-bjql4" Feb 18 15:11:35 crc kubenswrapper[4896]: I0218 15:11:35.915724 4896 reflector.go:368] Caches populated for *v1.Secret from object-"test-operator-kuttl-tests"/"keystone-keystone-dockercfg-vdf5c" Feb 18 15:11:35 crc kubenswrapper[4896]: I0218 15:11:35.916060 4896 reflector.go:368] Caches populated for *v1.Secret from object-"test-operator-kuttl-tests"/"keystone-config-data" Feb 18 15:11:35 crc kubenswrapper[4896]: I0218 15:11:35.916179 4896 reflector.go:368] Caches populated for *v1.Secret from object-"test-operator-kuttl-tests"/"keystone-scripts" Feb 18 15:11:35 crc kubenswrapper[4896]: I0218 15:11:35.916938 4896 reflector.go:368] Caches populated for *v1.Secret from object-"test-operator-kuttl-tests"/"keystone" Feb 18 15:11:35 crc kubenswrapper[4896]: I0218 15:11:35.924530 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["test-operator-kuttl-tests/keystone-7454c655dc-bjql4"] Feb 18 15:11:36 crc kubenswrapper[4896]: I0218 15:11:36.016708 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x27wf\" (UniqueName: \"kubernetes.io/projected/7b99a0f0-1581-40ba-b1a9-6311ac1be415-kube-api-access-x27wf\") pod \"keystone-7454c655dc-bjql4\" (UID: \"7b99a0f0-1581-40ba-b1a9-6311ac1be415\") " pod="test-operator-kuttl-tests/keystone-7454c655dc-bjql4" Feb 18 15:11:36 crc kubenswrapper[4896]: I0218 15:11:36.016773 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b99a0f0-1581-40ba-b1a9-6311ac1be415-scripts\") pod \"keystone-7454c655dc-bjql4\" (UID: \"7b99a0f0-1581-40ba-b1a9-6311ac1be415\") " pod="test-operator-kuttl-tests/keystone-7454c655dc-bjql4" Feb 18 15:11:36 crc kubenswrapper[4896]: I0218 15:11:36.016812 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7b99a0f0-1581-40ba-b1a9-6311ac1be415-credential-keys\") pod \"keystone-7454c655dc-bjql4\" (UID: \"7b99a0f0-1581-40ba-b1a9-6311ac1be415\") " pod="test-operator-kuttl-tests/keystone-7454c655dc-bjql4" Feb 18 15:11:36 crc kubenswrapper[4896]: I0218 15:11:36.016854 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b99a0f0-1581-40ba-b1a9-6311ac1be415-config-data\") pod \"keystone-7454c655dc-bjql4\" (UID: \"7b99a0f0-1581-40ba-b1a9-6311ac1be415\") " pod="test-operator-kuttl-tests/keystone-7454c655dc-bjql4" Feb 18 15:11:36 crc kubenswrapper[4896]: I0218 15:11:36.016882 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7b99a0f0-1581-40ba-b1a9-6311ac1be415-fernet-keys\") pod \"keystone-7454c655dc-bjql4\" (UID: \"7b99a0f0-1581-40ba-b1a9-6311ac1be415\") " pod="test-operator-kuttl-tests/keystone-7454c655dc-bjql4" Feb 18 15:11:36 crc kubenswrapper[4896]: I0218 15:11:36.117623 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b99a0f0-1581-40ba-b1a9-6311ac1be415-config-data\") pod \"keystone-7454c655dc-bjql4\" (UID: \"7b99a0f0-1581-40ba-b1a9-6311ac1be415\") " pod="test-operator-kuttl-tests/keystone-7454c655dc-bjql4" Feb 18 15:11:36 crc kubenswrapper[4896]: I0218 15:11:36.117688 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7b99a0f0-1581-40ba-b1a9-6311ac1be415-fernet-keys\") pod \"keystone-7454c655dc-bjql4\" (UID: \"7b99a0f0-1581-40ba-b1a9-6311ac1be415\") " pod="test-operator-kuttl-tests/keystone-7454c655dc-bjql4" Feb 18 15:11:36 crc kubenswrapper[4896]: I0218 15:11:36.117728 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x27wf\" (UniqueName: \"kubernetes.io/projected/7b99a0f0-1581-40ba-b1a9-6311ac1be415-kube-api-access-x27wf\") pod \"keystone-7454c655dc-bjql4\" (UID: \"7b99a0f0-1581-40ba-b1a9-6311ac1be415\") " pod="test-operator-kuttl-tests/keystone-7454c655dc-bjql4" Feb 18 15:11:36 crc kubenswrapper[4896]: I0218 15:11:36.117760 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b99a0f0-1581-40ba-b1a9-6311ac1be415-scripts\") pod \"keystone-7454c655dc-bjql4\" (UID: \"7b99a0f0-1581-40ba-b1a9-6311ac1be415\") " pod="test-operator-kuttl-tests/keystone-7454c655dc-bjql4" Feb 18 15:11:36 crc kubenswrapper[4896]: I0218 15:11:36.117801 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7b99a0f0-1581-40ba-b1a9-6311ac1be415-credential-keys\") pod \"keystone-7454c655dc-bjql4\" (UID: \"7b99a0f0-1581-40ba-b1a9-6311ac1be415\") " pod="test-operator-kuttl-tests/keystone-7454c655dc-bjql4" Feb 18 15:11:36 crc kubenswrapper[4896]: I0218 15:11:36.121347 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7b99a0f0-1581-40ba-b1a9-6311ac1be415-credential-keys\") pod \"keystone-7454c655dc-bjql4\" (UID: \"7b99a0f0-1581-40ba-b1a9-6311ac1be415\") " pod="test-operator-kuttl-tests/keystone-7454c655dc-bjql4" Feb 18 15:11:36 crc kubenswrapper[4896]: I0218 15:11:36.121554 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b99a0f0-1581-40ba-b1a9-6311ac1be415-scripts\") pod \"keystone-7454c655dc-bjql4\" (UID: \"7b99a0f0-1581-40ba-b1a9-6311ac1be415\") " pod="test-operator-kuttl-tests/keystone-7454c655dc-bjql4" Feb 18 15:11:36 crc kubenswrapper[4896]: I0218 15:11:36.121932 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7b99a0f0-1581-40ba-b1a9-6311ac1be415-fernet-keys\") pod \"keystone-7454c655dc-bjql4\" (UID: \"7b99a0f0-1581-40ba-b1a9-6311ac1be415\") " pod="test-operator-kuttl-tests/keystone-7454c655dc-bjql4" Feb 18 15:11:36 crc kubenswrapper[4896]: I0218 15:11:36.122916 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b99a0f0-1581-40ba-b1a9-6311ac1be415-config-data\") pod \"keystone-7454c655dc-bjql4\" (UID: \"7b99a0f0-1581-40ba-b1a9-6311ac1be415\") " pod="test-operator-kuttl-tests/keystone-7454c655dc-bjql4" Feb 18 15:11:36 crc kubenswrapper[4896]: I0218 15:11:36.136067 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x27wf\" (UniqueName: \"kubernetes.io/projected/7b99a0f0-1581-40ba-b1a9-6311ac1be415-kube-api-access-x27wf\") pod \"keystone-7454c655dc-bjql4\" (UID: \"7b99a0f0-1581-40ba-b1a9-6311ac1be415\") " pod="test-operator-kuttl-tests/keystone-7454c655dc-bjql4" Feb 18 15:11:36 crc kubenswrapper[4896]: I0218 15:11:36.236280 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="test-operator-kuttl-tests/keystone-7454c655dc-bjql4" Feb 18 15:11:37 crc kubenswrapper[4896]: I0218 15:11:37.417117 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["test-operator-kuttl-tests/keystone-7454c655dc-bjql4"] Feb 18 15:11:37 crc kubenswrapper[4896]: I0218 15:11:37.849103 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/keystone-7454c655dc-bjql4" event={"ID":"7b99a0f0-1581-40ba-b1a9-6311ac1be415","Type":"ContainerStarted","Data":"6bea6e1bbe6bca10309ee519852fd9b9529b2950cf7337e30af5d0faa6b47a52"} Feb 18 15:11:37 crc kubenswrapper[4896]: I0218 15:11:37.849451 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="test-operator-kuttl-tests/keystone-7454c655dc-bjql4" event={"ID":"7b99a0f0-1581-40ba-b1a9-6311ac1be415","Type":"ContainerStarted","Data":"22c43658666871e897e46c1950cf454d8705f443ac45f421ad740e23df8a4ede"} Feb 18 15:11:37 crc kubenswrapper[4896]: I0218 15:11:37.849571 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="test-operator-kuttl-tests/keystone-7454c655dc-bjql4" Feb 18 15:11:37 crc kubenswrapper[4896]: I0218 15:11:37.867575 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="test-operator-kuttl-tests/keystone-7454c655dc-bjql4" podStartSLOduration=2.867552995 podStartE2EDuration="2.867552995s" podCreationTimestamp="2026-02-18 15:11:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:11:37.862206109 +0000 UTC m=+996.791540322" watchObservedRunningTime="2026-02-18 15:11:37.867552995 +0000 UTC m=+996.796887208" Feb 18 15:12:08 crc kubenswrapper[4896]: I0218 15:12:08.031108 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="test-operator-kuttl-tests/keystone-7454c655dc-bjql4" Feb 18 15:13:04 crc kubenswrapper[4896]: I0218 15:13:04.075797 4896 patch_prober.go:28] interesting pod/machine-config-daemon-24rq5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:13:04 crc kubenswrapper[4896]: I0218 15:13:04.076347 4896 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:13:34 crc kubenswrapper[4896]: I0218 15:13:34.075299 4896 patch_prober.go:28] interesting pod/machine-config-daemon-24rq5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:13:34 crc kubenswrapper[4896]: I0218 15:13:34.075780 4896 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:14:04 crc kubenswrapper[4896]: I0218 15:14:04.076099 4896 patch_prober.go:28] interesting pod/machine-config-daemon-24rq5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:14:04 crc kubenswrapper[4896]: I0218 15:14:04.076706 4896 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:14:04 crc kubenswrapper[4896]: I0218 15:14:04.076759 4896 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" Feb 18 15:14:04 crc kubenswrapper[4896]: I0218 15:14:04.077505 4896 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1c40add7af533d9a1676f813ee5909cf520a8eec67c80f8dbfffb153fcc0d39e"} pod="openshift-machine-config-operator/machine-config-daemon-24rq5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 18 15:14:04 crc kubenswrapper[4896]: I0218 15:14:04.077549 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" containerID="cri-o://1c40add7af533d9a1676f813ee5909cf520a8eec67c80f8dbfffb153fcc0d39e" gracePeriod=600 Feb 18 15:14:04 crc kubenswrapper[4896]: I0218 15:14:04.739428 4896 generic.go:334] "Generic (PLEG): container finished" podID="8503734a-8111-4fab-beda-1f0f3b59615b" containerID="1c40add7af533d9a1676f813ee5909cf520a8eec67c80f8dbfffb153fcc0d39e" exitCode=0 Feb 18 15:14:04 crc kubenswrapper[4896]: I0218 15:14:04.739449 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" event={"ID":"8503734a-8111-4fab-beda-1f0f3b59615b","Type":"ContainerDied","Data":"1c40add7af533d9a1676f813ee5909cf520a8eec67c80f8dbfffb153fcc0d39e"} Feb 18 15:14:04 crc kubenswrapper[4896]: I0218 15:14:04.740127 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" event={"ID":"8503734a-8111-4fab-beda-1f0f3b59615b","Type":"ContainerStarted","Data":"6f33e818698510245a241c7b8b0945082ac2ca4f2f7697e38796ef117ee06b80"} Feb 18 15:14:04 crc kubenswrapper[4896]: I0218 15:14:04.740166 4896 scope.go:117] "RemoveContainer" containerID="763c989180a0be50e9c5a0c490b89d9b8795e413a592c48ba235f513fd0066bc" Feb 18 15:15:00 crc kubenswrapper[4896]: I0218 15:15:00.151399 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523795-5j2n9"] Feb 18 15:15:00 crc kubenswrapper[4896]: I0218 15:15:00.153263 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-5j2n9" Feb 18 15:15:00 crc kubenswrapper[4896]: I0218 15:15:00.155395 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 18 15:15:00 crc kubenswrapper[4896]: I0218 15:15:00.159489 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523795-5j2n9"] Feb 18 15:15:00 crc kubenswrapper[4896]: I0218 15:15:00.161676 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 18 15:15:00 crc kubenswrapper[4896]: I0218 15:15:00.316165 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1272dfa4-8d81-43ac-8312-b243267500e5-secret-volume\") pod \"collect-profiles-29523795-5j2n9\" (UID: \"1272dfa4-8d81-43ac-8312-b243267500e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-5j2n9" Feb 18 15:15:00 crc kubenswrapper[4896]: I0218 15:15:00.316222 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1272dfa4-8d81-43ac-8312-b243267500e5-config-volume\") pod \"collect-profiles-29523795-5j2n9\" (UID: \"1272dfa4-8d81-43ac-8312-b243267500e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-5j2n9" Feb 18 15:15:00 crc kubenswrapper[4896]: I0218 15:15:00.316328 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzp2g\" (UniqueName: \"kubernetes.io/projected/1272dfa4-8d81-43ac-8312-b243267500e5-kube-api-access-jzp2g\") pod \"collect-profiles-29523795-5j2n9\" (UID: \"1272dfa4-8d81-43ac-8312-b243267500e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-5j2n9" Feb 18 15:15:00 crc kubenswrapper[4896]: I0218 15:15:00.417781 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1272dfa4-8d81-43ac-8312-b243267500e5-secret-volume\") pod \"collect-profiles-29523795-5j2n9\" (UID: \"1272dfa4-8d81-43ac-8312-b243267500e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-5j2n9" Feb 18 15:15:00 crc kubenswrapper[4896]: I0218 15:15:00.417826 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1272dfa4-8d81-43ac-8312-b243267500e5-config-volume\") pod \"collect-profiles-29523795-5j2n9\" (UID: \"1272dfa4-8d81-43ac-8312-b243267500e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-5j2n9" Feb 18 15:15:00 crc kubenswrapper[4896]: I0218 15:15:00.417888 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzp2g\" (UniqueName: \"kubernetes.io/projected/1272dfa4-8d81-43ac-8312-b243267500e5-kube-api-access-jzp2g\") pod \"collect-profiles-29523795-5j2n9\" (UID: \"1272dfa4-8d81-43ac-8312-b243267500e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-5j2n9" Feb 18 15:15:00 crc kubenswrapper[4896]: I0218 15:15:00.418800 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1272dfa4-8d81-43ac-8312-b243267500e5-config-volume\") pod \"collect-profiles-29523795-5j2n9\" (UID: \"1272dfa4-8d81-43ac-8312-b243267500e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-5j2n9" Feb 18 15:15:00 crc kubenswrapper[4896]: I0218 15:15:00.428281 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1272dfa4-8d81-43ac-8312-b243267500e5-secret-volume\") pod \"collect-profiles-29523795-5j2n9\" (UID: \"1272dfa4-8d81-43ac-8312-b243267500e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-5j2n9" Feb 18 15:15:00 crc kubenswrapper[4896]: I0218 15:15:00.442423 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzp2g\" (UniqueName: \"kubernetes.io/projected/1272dfa4-8d81-43ac-8312-b243267500e5-kube-api-access-jzp2g\") pod \"collect-profiles-29523795-5j2n9\" (UID: \"1272dfa4-8d81-43ac-8312-b243267500e5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-5j2n9" Feb 18 15:15:00 crc kubenswrapper[4896]: I0218 15:15:00.477196 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-5j2n9" Feb 18 15:15:00 crc kubenswrapper[4896]: I0218 15:15:00.888156 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29523795-5j2n9"] Feb 18 15:15:01 crc kubenswrapper[4896]: I0218 15:15:01.119870 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-5j2n9" event={"ID":"1272dfa4-8d81-43ac-8312-b243267500e5","Type":"ContainerStarted","Data":"403c2a578ae390a42115b2d49e3fea6ec310096457c4b4dc12b33e77bd1c1aff"} Feb 18 15:15:01 crc kubenswrapper[4896]: I0218 15:15:01.120002 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-5j2n9" event={"ID":"1272dfa4-8d81-43ac-8312-b243267500e5","Type":"ContainerStarted","Data":"744cfee0f48ae1e58e78fbf4d3ef83cd8f1cff46d9e504be055d7f55fb68f94e"} Feb 18 15:15:01 crc kubenswrapper[4896]: I0218 15:15:01.136032 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-5j2n9" podStartSLOduration=1.136016954 podStartE2EDuration="1.136016954s" podCreationTimestamp="2026-02-18 15:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-18 15:15:01.13396911 +0000 UTC m=+1200.063303313" watchObservedRunningTime="2026-02-18 15:15:01.136016954 +0000 UTC m=+1200.065351167" Feb 18 15:15:02 crc kubenswrapper[4896]: I0218 15:15:02.128016 4896 generic.go:334] "Generic (PLEG): container finished" podID="1272dfa4-8d81-43ac-8312-b243267500e5" containerID="403c2a578ae390a42115b2d49e3fea6ec310096457c4b4dc12b33e77bd1c1aff" exitCode=0 Feb 18 15:15:02 crc kubenswrapper[4896]: I0218 15:15:02.128060 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-5j2n9" event={"ID":"1272dfa4-8d81-43ac-8312-b243267500e5","Type":"ContainerDied","Data":"403c2a578ae390a42115b2d49e3fea6ec310096457c4b4dc12b33e77bd1c1aff"} Feb 18 15:15:03 crc kubenswrapper[4896]: I0218 15:15:03.365814 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-5j2n9" Feb 18 15:15:03 crc kubenswrapper[4896]: I0218 15:15:03.457662 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1272dfa4-8d81-43ac-8312-b243267500e5-config-volume\") pod \"1272dfa4-8d81-43ac-8312-b243267500e5\" (UID: \"1272dfa4-8d81-43ac-8312-b243267500e5\") " Feb 18 15:15:03 crc kubenswrapper[4896]: I0218 15:15:03.457825 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1272dfa4-8d81-43ac-8312-b243267500e5-secret-volume\") pod \"1272dfa4-8d81-43ac-8312-b243267500e5\" (UID: \"1272dfa4-8d81-43ac-8312-b243267500e5\") " Feb 18 15:15:03 crc kubenswrapper[4896]: I0218 15:15:03.458548 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzp2g\" (UniqueName: \"kubernetes.io/projected/1272dfa4-8d81-43ac-8312-b243267500e5-kube-api-access-jzp2g\") pod \"1272dfa4-8d81-43ac-8312-b243267500e5\" (UID: \"1272dfa4-8d81-43ac-8312-b243267500e5\") " Feb 18 15:15:03 crc kubenswrapper[4896]: I0218 15:15:03.458586 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1272dfa4-8d81-43ac-8312-b243267500e5-config-volume" (OuterVolumeSpecName: "config-volume") pod "1272dfa4-8d81-43ac-8312-b243267500e5" (UID: "1272dfa4-8d81-43ac-8312-b243267500e5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 18 15:15:03 crc kubenswrapper[4896]: I0218 15:15:03.459113 4896 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1272dfa4-8d81-43ac-8312-b243267500e5-config-volume\") on node \"crc\" DevicePath \"\"" Feb 18 15:15:03 crc kubenswrapper[4896]: I0218 15:15:03.465099 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1272dfa4-8d81-43ac-8312-b243267500e5-kube-api-access-jzp2g" (OuterVolumeSpecName: "kube-api-access-jzp2g") pod "1272dfa4-8d81-43ac-8312-b243267500e5" (UID: "1272dfa4-8d81-43ac-8312-b243267500e5"). InnerVolumeSpecName "kube-api-access-jzp2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:15:03 crc kubenswrapper[4896]: I0218 15:15:03.466009 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1272dfa4-8d81-43ac-8312-b243267500e5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1272dfa4-8d81-43ac-8312-b243267500e5" (UID: "1272dfa4-8d81-43ac-8312-b243267500e5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 18 15:15:03 crc kubenswrapper[4896]: I0218 15:15:03.560999 4896 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1272dfa4-8d81-43ac-8312-b243267500e5-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 18 15:15:03 crc kubenswrapper[4896]: I0218 15:15:03.561041 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzp2g\" (UniqueName: \"kubernetes.io/projected/1272dfa4-8d81-43ac-8312-b243267500e5-kube-api-access-jzp2g\") on node \"crc\" DevicePath \"\"" Feb 18 15:15:04 crc kubenswrapper[4896]: I0218 15:15:04.141703 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-5j2n9" event={"ID":"1272dfa4-8d81-43ac-8312-b243267500e5","Type":"ContainerDied","Data":"744cfee0f48ae1e58e78fbf4d3ef83cd8f1cff46d9e504be055d7f55fb68f94e"} Feb 18 15:15:04 crc kubenswrapper[4896]: I0218 15:15:04.141745 4896 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="744cfee0f48ae1e58e78fbf4d3ef83cd8f1cff46d9e504be055d7f55fb68f94e" Feb 18 15:15:04 crc kubenswrapper[4896]: I0218 15:15:04.141791 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29523795-5j2n9" Feb 18 15:16:04 crc kubenswrapper[4896]: I0218 15:16:04.075877 4896 patch_prober.go:28] interesting pod/machine-config-daemon-24rq5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:16:04 crc kubenswrapper[4896]: I0218 15:16:04.076470 4896 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:16:34 crc kubenswrapper[4896]: I0218 15:16:34.076075 4896 patch_prober.go:28] interesting pod/machine-config-daemon-24rq5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:16:34 crc kubenswrapper[4896]: I0218 15:16:34.076627 4896 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:17:04 crc kubenswrapper[4896]: I0218 15:17:04.075557 4896 patch_prober.go:28] interesting pod/machine-config-daemon-24rq5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:17:04 crc kubenswrapper[4896]: I0218 15:17:04.077127 4896 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:17:04 crc kubenswrapper[4896]: I0218 15:17:04.077258 4896 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" Feb 18 15:17:04 crc kubenswrapper[4896]: I0218 15:17:04.077930 4896 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6f33e818698510245a241c7b8b0945082ac2ca4f2f7697e38796ef117ee06b80"} pod="openshift-machine-config-operator/machine-config-daemon-24rq5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 18 15:17:04 crc kubenswrapper[4896]: I0218 15:17:04.078070 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" containerID="cri-o://6f33e818698510245a241c7b8b0945082ac2ca4f2f7697e38796ef117ee06b80" gracePeriod=600 Feb 18 15:17:04 crc kubenswrapper[4896]: I0218 15:17:04.887746 4896 generic.go:334] "Generic (PLEG): container finished" podID="8503734a-8111-4fab-beda-1f0f3b59615b" containerID="6f33e818698510245a241c7b8b0945082ac2ca4f2f7697e38796ef117ee06b80" exitCode=0 Feb 18 15:17:04 crc kubenswrapper[4896]: I0218 15:17:04.887863 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" event={"ID":"8503734a-8111-4fab-beda-1f0f3b59615b","Type":"ContainerDied","Data":"6f33e818698510245a241c7b8b0945082ac2ca4f2f7697e38796ef117ee06b80"} Feb 18 15:17:04 crc kubenswrapper[4896]: I0218 15:17:04.888094 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" event={"ID":"8503734a-8111-4fab-beda-1f0f3b59615b","Type":"ContainerStarted","Data":"7cbf684d610fed53f6fdd655adb1877f0df69baa4d3a58b561f973625a82e83c"} Feb 18 15:17:04 crc kubenswrapper[4896]: I0218 15:17:04.888116 4896 scope.go:117] "RemoveContainer" containerID="1c40add7af533d9a1676f813ee5909cf520a8eec67c80f8dbfffb153fcc0d39e" Feb 18 15:19:04 crc kubenswrapper[4896]: I0218 15:19:04.075639 4896 patch_prober.go:28] interesting pod/machine-config-daemon-24rq5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:19:04 crc kubenswrapper[4896]: I0218 15:19:04.076121 4896 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:19:33 crc kubenswrapper[4896]: I0218 15:19:33.062420 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["test-operator-kuttl-tests/root-account-create-update-45cql"] Feb 18 15:19:33 crc kubenswrapper[4896]: I0218 15:19:33.073445 4896 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["test-operator-kuttl-tests/root-account-create-update-45cql"] Feb 18 15:19:33 crc kubenswrapper[4896]: I0218 15:19:33.935386 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe758f2a-b576-4713-9803-4d34178cfebf" path="/var/lib/kubelet/pods/fe758f2a-b576-4713-9803-4d34178cfebf/volumes" Feb 18 15:19:34 crc kubenswrapper[4896]: I0218 15:19:34.075534 4896 patch_prober.go:28] interesting pod/machine-config-daemon-24rq5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:19:34 crc kubenswrapper[4896]: I0218 15:19:34.075611 4896 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:20:01 crc kubenswrapper[4896]: I0218 15:20:01.969744 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-86dmm/must-gather-7jbtf"] Feb 18 15:20:01 crc kubenswrapper[4896]: E0218 15:20:01.970573 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1272dfa4-8d81-43ac-8312-b243267500e5" containerName="collect-profiles" Feb 18 15:20:01 crc kubenswrapper[4896]: I0218 15:20:01.970591 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="1272dfa4-8d81-43ac-8312-b243267500e5" containerName="collect-profiles" Feb 18 15:20:01 crc kubenswrapper[4896]: I0218 15:20:01.970753 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="1272dfa4-8d81-43ac-8312-b243267500e5" containerName="collect-profiles" Feb 18 15:20:01 crc kubenswrapper[4896]: I0218 15:20:01.971621 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-86dmm/must-gather-7jbtf" Feb 18 15:20:01 crc kubenswrapper[4896]: I0218 15:20:01.973718 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-86dmm"/"openshift-service-ca.crt" Feb 18 15:20:01 crc kubenswrapper[4896]: I0218 15:20:01.974338 4896 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-86dmm"/"default-dockercfg-mbnpc" Feb 18 15:20:01 crc kubenswrapper[4896]: I0218 15:20:01.974378 4896 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-86dmm"/"kube-root-ca.crt" Feb 18 15:20:01 crc kubenswrapper[4896]: I0218 15:20:01.996978 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-86dmm/must-gather-7jbtf"] Feb 18 15:20:02 crc kubenswrapper[4896]: I0218 15:20:02.169355 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5hbz\" (UniqueName: \"kubernetes.io/projected/7288ac23-5347-46d8-b5fc-ba214158bf26-kube-api-access-q5hbz\") pod \"must-gather-7jbtf\" (UID: \"7288ac23-5347-46d8-b5fc-ba214158bf26\") " pod="openshift-must-gather-86dmm/must-gather-7jbtf" Feb 18 15:20:02 crc kubenswrapper[4896]: I0218 15:20:02.169457 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7288ac23-5347-46d8-b5fc-ba214158bf26-must-gather-output\") pod \"must-gather-7jbtf\" (UID: \"7288ac23-5347-46d8-b5fc-ba214158bf26\") " pod="openshift-must-gather-86dmm/must-gather-7jbtf" Feb 18 15:20:02 crc kubenswrapper[4896]: I0218 15:20:02.270706 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5hbz\" (UniqueName: \"kubernetes.io/projected/7288ac23-5347-46d8-b5fc-ba214158bf26-kube-api-access-q5hbz\") pod \"must-gather-7jbtf\" (UID: \"7288ac23-5347-46d8-b5fc-ba214158bf26\") " pod="openshift-must-gather-86dmm/must-gather-7jbtf" Feb 18 15:20:02 crc kubenswrapper[4896]: I0218 15:20:02.270758 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7288ac23-5347-46d8-b5fc-ba214158bf26-must-gather-output\") pod \"must-gather-7jbtf\" (UID: \"7288ac23-5347-46d8-b5fc-ba214158bf26\") " pod="openshift-must-gather-86dmm/must-gather-7jbtf" Feb 18 15:20:02 crc kubenswrapper[4896]: I0218 15:20:02.271174 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7288ac23-5347-46d8-b5fc-ba214158bf26-must-gather-output\") pod \"must-gather-7jbtf\" (UID: \"7288ac23-5347-46d8-b5fc-ba214158bf26\") " pod="openshift-must-gather-86dmm/must-gather-7jbtf" Feb 18 15:20:02 crc kubenswrapper[4896]: I0218 15:20:02.286870 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5hbz\" (UniqueName: \"kubernetes.io/projected/7288ac23-5347-46d8-b5fc-ba214158bf26-kube-api-access-q5hbz\") pod \"must-gather-7jbtf\" (UID: \"7288ac23-5347-46d8-b5fc-ba214158bf26\") " pod="openshift-must-gather-86dmm/must-gather-7jbtf" Feb 18 15:20:02 crc kubenswrapper[4896]: I0218 15:20:02.288697 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-86dmm/must-gather-7jbtf" Feb 18 15:20:02 crc kubenswrapper[4896]: I0218 15:20:02.377726 4896 scope.go:117] "RemoveContainer" containerID="f3d17a83509b51dad78c25b94c855ae20d8fb865435518f3a227dd89df01c39c" Feb 18 15:20:02 crc kubenswrapper[4896]: I0218 15:20:02.523056 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-86dmm/must-gather-7jbtf"] Feb 18 15:20:02 crc kubenswrapper[4896]: I0218 15:20:02.532692 4896 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 18 15:20:03 crc kubenswrapper[4896]: I0218 15:20:03.067827 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-86dmm/must-gather-7jbtf" event={"ID":"7288ac23-5347-46d8-b5fc-ba214158bf26","Type":"ContainerStarted","Data":"5393bc0a7ac5b8da3b13f693830c9909a48512ace3ec57e0d360da8842e0d540"} Feb 18 15:20:04 crc kubenswrapper[4896]: I0218 15:20:04.076191 4896 patch_prober.go:28] interesting pod/machine-config-daemon-24rq5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:20:04 crc kubenswrapper[4896]: I0218 15:20:04.076560 4896 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:20:04 crc kubenswrapper[4896]: I0218 15:20:04.076608 4896 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" Feb 18 15:20:04 crc kubenswrapper[4896]: I0218 15:20:04.077302 4896 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7cbf684d610fed53f6fdd655adb1877f0df69baa4d3a58b561f973625a82e83c"} pod="openshift-machine-config-operator/machine-config-daemon-24rq5" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 18 15:20:04 crc kubenswrapper[4896]: I0218 15:20:04.077365 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" containerID="cri-o://7cbf684d610fed53f6fdd655adb1877f0df69baa4d3a58b561f973625a82e83c" gracePeriod=600 Feb 18 15:20:04 crc kubenswrapper[4896]: E0218 15:20:04.219395 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-24rq5_openshift-machine-config-operator(8503734a-8111-4fab-beda-1f0f3b59615b)\"" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" Feb 18 15:20:05 crc kubenswrapper[4896]: I0218 15:20:05.082820 4896 generic.go:334] "Generic (PLEG): container finished" podID="8503734a-8111-4fab-beda-1f0f3b59615b" containerID="7cbf684d610fed53f6fdd655adb1877f0df69baa4d3a58b561f973625a82e83c" exitCode=0 Feb 18 15:20:05 crc kubenswrapper[4896]: I0218 15:20:05.082881 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" event={"ID":"8503734a-8111-4fab-beda-1f0f3b59615b","Type":"ContainerDied","Data":"7cbf684d610fed53f6fdd655adb1877f0df69baa4d3a58b561f973625a82e83c"} Feb 18 15:20:05 crc kubenswrapper[4896]: I0218 15:20:05.082918 4896 scope.go:117] "RemoveContainer" containerID="6f33e818698510245a241c7b8b0945082ac2ca4f2f7697e38796ef117ee06b80" Feb 18 15:20:05 crc kubenswrapper[4896]: I0218 15:20:05.083365 4896 scope.go:117] "RemoveContainer" containerID="7cbf684d610fed53f6fdd655adb1877f0df69baa4d3a58b561f973625a82e83c" Feb 18 15:20:05 crc kubenswrapper[4896]: E0218 15:20:05.083547 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-24rq5_openshift-machine-config-operator(8503734a-8111-4fab-beda-1f0f3b59615b)\"" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" Feb 18 15:20:10 crc kubenswrapper[4896]: I0218 15:20:10.135981 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-86dmm/must-gather-7jbtf" event={"ID":"7288ac23-5347-46d8-b5fc-ba214158bf26","Type":"ContainerStarted","Data":"b89b86a45df21b5b2354f3b379c54efaf60c2dd0114aa863916974f0e297a6fd"} Feb 18 15:20:11 crc kubenswrapper[4896]: I0218 15:20:11.145755 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-86dmm/must-gather-7jbtf" event={"ID":"7288ac23-5347-46d8-b5fc-ba214158bf26","Type":"ContainerStarted","Data":"22108ab6b903f1f8329b2fb04c9dba418c8e363018e0dbd2e709907b76d8bee9"} Feb 18 15:20:11 crc kubenswrapper[4896]: I0218 15:20:11.162993 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-86dmm/must-gather-7jbtf" podStartSLOduration=2.829597557 podStartE2EDuration="10.162974073s" podCreationTimestamp="2026-02-18 15:20:01 +0000 UTC" firstStartedPulling="2026-02-18 15:20:02.532415695 +0000 UTC m=+1501.461749908" lastFinishedPulling="2026-02-18 15:20:09.865792211 +0000 UTC m=+1508.795126424" observedRunningTime="2026-02-18 15:20:11.157962502 +0000 UTC m=+1510.087296715" watchObservedRunningTime="2026-02-18 15:20:11.162974073 +0000 UTC m=+1510.092308286" Feb 18 15:20:17 crc kubenswrapper[4896]: I0218 15:20:17.927996 4896 scope.go:117] "RemoveContainer" containerID="7cbf684d610fed53f6fdd655adb1877f0df69baa4d3a58b561f973625a82e83c" Feb 18 15:20:17 crc kubenswrapper[4896]: E0218 15:20:17.928491 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-24rq5_openshift-machine-config-operator(8503734a-8111-4fab-beda-1f0f3b59615b)\"" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" Feb 18 15:20:28 crc kubenswrapper[4896]: I0218 15:20:28.927777 4896 scope.go:117] "RemoveContainer" containerID="7cbf684d610fed53f6fdd655adb1877f0df69baa4d3a58b561f973625a82e83c" Feb 18 15:20:28 crc kubenswrapper[4896]: E0218 15:20:28.928506 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-24rq5_openshift-machine-config-operator(8503734a-8111-4fab-beda-1f0f3b59615b)\"" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" Feb 18 15:20:41 crc kubenswrapper[4896]: I0218 15:20:41.223396 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qst58"] Feb 18 15:20:41 crc kubenswrapper[4896]: I0218 15:20:41.227121 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qst58" Feb 18 15:20:41 crc kubenswrapper[4896]: I0218 15:20:41.265769 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qst58"] Feb 18 15:20:41 crc kubenswrapper[4896]: I0218 15:20:41.339955 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c51c69de-45a2-44ad-834d-47cb43e328b9-utilities\") pod \"redhat-operators-qst58\" (UID: \"c51c69de-45a2-44ad-834d-47cb43e328b9\") " pod="openshift-marketplace/redhat-operators-qst58" Feb 18 15:20:41 crc kubenswrapper[4896]: I0218 15:20:41.340080 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9qtf\" (UniqueName: \"kubernetes.io/projected/c51c69de-45a2-44ad-834d-47cb43e328b9-kube-api-access-z9qtf\") pod \"redhat-operators-qst58\" (UID: \"c51c69de-45a2-44ad-834d-47cb43e328b9\") " pod="openshift-marketplace/redhat-operators-qst58" Feb 18 15:20:41 crc kubenswrapper[4896]: I0218 15:20:41.340126 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c51c69de-45a2-44ad-834d-47cb43e328b9-catalog-content\") pod \"redhat-operators-qst58\" (UID: \"c51c69de-45a2-44ad-834d-47cb43e328b9\") " pod="openshift-marketplace/redhat-operators-qst58" Feb 18 15:20:41 crc kubenswrapper[4896]: I0218 15:20:41.441120 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9qtf\" (UniqueName: \"kubernetes.io/projected/c51c69de-45a2-44ad-834d-47cb43e328b9-kube-api-access-z9qtf\") pod \"redhat-operators-qst58\" (UID: \"c51c69de-45a2-44ad-834d-47cb43e328b9\") " pod="openshift-marketplace/redhat-operators-qst58" Feb 18 15:20:41 crc kubenswrapper[4896]: I0218 15:20:41.441197 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c51c69de-45a2-44ad-834d-47cb43e328b9-catalog-content\") pod \"redhat-operators-qst58\" (UID: \"c51c69de-45a2-44ad-834d-47cb43e328b9\") " pod="openshift-marketplace/redhat-operators-qst58" Feb 18 15:20:41 crc kubenswrapper[4896]: I0218 15:20:41.441254 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c51c69de-45a2-44ad-834d-47cb43e328b9-utilities\") pod \"redhat-operators-qst58\" (UID: \"c51c69de-45a2-44ad-834d-47cb43e328b9\") " pod="openshift-marketplace/redhat-operators-qst58" Feb 18 15:20:41 crc kubenswrapper[4896]: I0218 15:20:41.441821 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c51c69de-45a2-44ad-834d-47cb43e328b9-catalog-content\") pod \"redhat-operators-qst58\" (UID: \"c51c69de-45a2-44ad-834d-47cb43e328b9\") " pod="openshift-marketplace/redhat-operators-qst58" Feb 18 15:20:41 crc kubenswrapper[4896]: I0218 15:20:41.441869 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c51c69de-45a2-44ad-834d-47cb43e328b9-utilities\") pod \"redhat-operators-qst58\" (UID: \"c51c69de-45a2-44ad-834d-47cb43e328b9\") " pod="openshift-marketplace/redhat-operators-qst58" Feb 18 15:20:41 crc kubenswrapper[4896]: I0218 15:20:41.462736 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9qtf\" (UniqueName: \"kubernetes.io/projected/c51c69de-45a2-44ad-834d-47cb43e328b9-kube-api-access-z9qtf\") pod \"redhat-operators-qst58\" (UID: \"c51c69de-45a2-44ad-834d-47cb43e328b9\") " pod="openshift-marketplace/redhat-operators-qst58" Feb 18 15:20:41 crc kubenswrapper[4896]: I0218 15:20:41.601021 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qst58" Feb 18 15:20:41 crc kubenswrapper[4896]: I0218 15:20:41.840846 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qst58"] Feb 18 15:20:42 crc kubenswrapper[4896]: I0218 15:20:42.317786 4896 generic.go:334] "Generic (PLEG): container finished" podID="c51c69de-45a2-44ad-834d-47cb43e328b9" containerID="c3fc429252aa3569b6127506375a1b2a9142768aa776246dc0b4c2fd72709e3c" exitCode=0 Feb 18 15:20:42 crc kubenswrapper[4896]: I0218 15:20:42.317833 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qst58" event={"ID":"c51c69de-45a2-44ad-834d-47cb43e328b9","Type":"ContainerDied","Data":"c3fc429252aa3569b6127506375a1b2a9142768aa776246dc0b4c2fd72709e3c"} Feb 18 15:20:42 crc kubenswrapper[4896]: I0218 15:20:42.317887 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qst58" event={"ID":"c51c69de-45a2-44ad-834d-47cb43e328b9","Type":"ContainerStarted","Data":"027192e2a32397767c95ec863374f7d5d0df623d534bab229c229bd4ec0185bb"} Feb 18 15:20:42 crc kubenswrapper[4896]: I0218 15:20:42.928142 4896 scope.go:117] "RemoveContainer" containerID="7cbf684d610fed53f6fdd655adb1877f0df69baa4d3a58b561f973625a82e83c" Feb 18 15:20:42 crc kubenswrapper[4896]: E0218 15:20:42.928608 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-24rq5_openshift-machine-config-operator(8503734a-8111-4fab-beda-1f0f3b59615b)\"" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" Feb 18 15:20:43 crc kubenswrapper[4896]: I0218 15:20:43.325785 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qst58" event={"ID":"c51c69de-45a2-44ad-834d-47cb43e328b9","Type":"ContainerStarted","Data":"b95e3b8fb24b92809bfecca896de913d5bd1df427c5014c6bcddc91d0890557d"} Feb 18 15:20:44 crc kubenswrapper[4896]: I0218 15:20:44.332387 4896 generic.go:334] "Generic (PLEG): container finished" podID="c51c69de-45a2-44ad-834d-47cb43e328b9" containerID="b95e3b8fb24b92809bfecca896de913d5bd1df427c5014c6bcddc91d0890557d" exitCode=0 Feb 18 15:20:44 crc kubenswrapper[4896]: I0218 15:20:44.332686 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qst58" event={"ID":"c51c69de-45a2-44ad-834d-47cb43e328b9","Type":"ContainerDied","Data":"b95e3b8fb24b92809bfecca896de913d5bd1df427c5014c6bcddc91d0890557d"} Feb 18 15:20:45 crc kubenswrapper[4896]: I0218 15:20:45.338948 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qst58" event={"ID":"c51c69de-45a2-44ad-834d-47cb43e328b9","Type":"ContainerStarted","Data":"cd2cd75827c03c15067dd3f71cf1dd99348ca66610d41c21d341d5c7cd1f5a78"} Feb 18 15:20:45 crc kubenswrapper[4896]: I0218 15:20:45.360674 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qst58" podStartSLOduration=1.941182009 podStartE2EDuration="4.360660826s" podCreationTimestamp="2026-02-18 15:20:41 +0000 UTC" firstStartedPulling="2026-02-18 15:20:42.319708777 +0000 UTC m=+1541.249042990" lastFinishedPulling="2026-02-18 15:20:44.739187594 +0000 UTC m=+1543.668521807" observedRunningTime="2026-02-18 15:20:45.358452834 +0000 UTC m=+1544.287787047" watchObservedRunningTime="2026-02-18 15:20:45.360660826 +0000 UTC m=+1544.289995039" Feb 18 15:20:48 crc kubenswrapper[4896]: I0218 15:20:48.670760 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb_688b59fd-cdfb-421e-beff-abf0651ab5f2/util/0.log" Feb 18 15:20:48 crc kubenswrapper[4896]: I0218 15:20:48.832225 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb_688b59fd-cdfb-421e-beff-abf0651ab5f2/util/0.log" Feb 18 15:20:48 crc kubenswrapper[4896]: I0218 15:20:48.867517 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb_688b59fd-cdfb-421e-beff-abf0651ab5f2/pull/0.log" Feb 18 15:20:48 crc kubenswrapper[4896]: I0218 15:20:48.904356 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb_688b59fd-cdfb-421e-beff-abf0651ab5f2/pull/0.log" Feb 18 15:20:49 crc kubenswrapper[4896]: I0218 15:20:49.146058 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb_688b59fd-cdfb-421e-beff-abf0651ab5f2/extract/0.log" Feb 18 15:20:49 crc kubenswrapper[4896]: I0218 15:20:49.270737 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb_688b59fd-cdfb-421e-beff-abf0651ab5f2/util/0.log" Feb 18 15:20:49 crc kubenswrapper[4896]: I0218 15:20:49.271199 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_414e4b97d51928ab5198aefe9c55ad0e5126b10e101f1abf7c39f91e0682xpb_688b59fd-cdfb-421e-beff-abf0651ab5f2/pull/0.log" Feb 18 15:20:49 crc kubenswrapper[4896]: I0218 15:20:49.310400 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt_87e452f6-c88e-4997-81e9-cab5976ab979/util/0.log" Feb 18 15:20:49 crc kubenswrapper[4896]: I0218 15:20:49.563397 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt_87e452f6-c88e-4997-81e9-cab5976ab979/pull/0.log" Feb 18 15:20:49 crc kubenswrapper[4896]: I0218 15:20:49.586067 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt_87e452f6-c88e-4997-81e9-cab5976ab979/pull/0.log" Feb 18 15:20:49 crc kubenswrapper[4896]: I0218 15:20:49.601657 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt_87e452f6-c88e-4997-81e9-cab5976ab979/util/0.log" Feb 18 15:20:49 crc kubenswrapper[4896]: I0218 15:20:49.736581 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt_87e452f6-c88e-4997-81e9-cab5976ab979/util/0.log" Feb 18 15:20:49 crc kubenswrapper[4896]: I0218 15:20:49.787258 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt_87e452f6-c88e-4997-81e9-cab5976ab979/extract/0.log" Feb 18 15:20:49 crc kubenswrapper[4896]: I0218 15:20:49.821128 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_4d4a9dcdacca526e9e6f5178499d735b8b4c6fd7e962363bed41c17b9bk7zdt_87e452f6-c88e-4997-81e9-cab5976ab979/pull/0.log" Feb 18 15:20:49 crc kubenswrapper[4896]: I0218 15:20:49.988447 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp_c6747e4d-4e79-4199-942b-8354cbafa962/util/0.log" Feb 18 15:20:50 crc kubenswrapper[4896]: I0218 15:20:50.228827 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp_c6747e4d-4e79-4199-942b-8354cbafa962/util/0.log" Feb 18 15:20:50 crc kubenswrapper[4896]: I0218 15:20:50.229427 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp_c6747e4d-4e79-4199-942b-8354cbafa962/pull/0.log" Feb 18 15:20:50 crc kubenswrapper[4896]: I0218 15:20:50.254789 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp_c6747e4d-4e79-4199-942b-8354cbafa962/pull/0.log" Feb 18 15:20:50 crc kubenswrapper[4896]: I0218 15:20:50.467471 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp_c6747e4d-4e79-4199-942b-8354cbafa962/util/0.log" Feb 18 15:20:50 crc kubenswrapper[4896]: I0218 15:20:50.472829 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp_c6747e4d-4e79-4199-942b-8354cbafa962/pull/0.log" Feb 18 15:20:50 crc kubenswrapper[4896]: I0218 15:20:50.508156 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e5907s6wp_c6747e4d-4e79-4199-942b-8354cbafa962/extract/0.log" Feb 18 15:20:50 crc kubenswrapper[4896]: I0218 15:20:50.662282 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f_75e1f6d2-45d3-4805-9382-32d8a89ece0a/util/0.log" Feb 18 15:20:50 crc kubenswrapper[4896]: I0218 15:20:50.874610 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f_75e1f6d2-45d3-4805-9382-32d8a89ece0a/pull/0.log" Feb 18 15:20:50 crc kubenswrapper[4896]: I0218 15:20:50.908572 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f_75e1f6d2-45d3-4805-9382-32d8a89ece0a/pull/0.log" Feb 18 15:20:50 crc kubenswrapper[4896]: I0218 15:20:50.921692 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f_75e1f6d2-45d3-4805-9382-32d8a89ece0a/util/0.log" Feb 18 15:20:51 crc kubenswrapper[4896]: I0218 15:20:51.094901 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f_75e1f6d2-45d3-4805-9382-32d8a89ece0a/extract/0.log" Feb 18 15:20:51 crc kubenswrapper[4896]: I0218 15:20:51.125885 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f_75e1f6d2-45d3-4805-9382-32d8a89ece0a/pull/0.log" Feb 18 15:20:51 crc kubenswrapper[4896]: I0218 15:20:51.161649 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_97c0565e024cc42acedf7d327ee11bcdec84a0c0e4ddd546647cdbfd63rjk4f_75e1f6d2-45d3-4805-9382-32d8a89ece0a/util/0.log" Feb 18 15:20:51 crc kubenswrapper[4896]: I0218 15:20:51.302807 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5d86b96dc9-dhl2n_bf5a2769-d00a-4493-825f-006f6607e79b/manager/0.log" Feb 18 15:20:51 crc kubenswrapper[4896]: I0218 15:20:51.338106 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-index-46bnc_51243d7d-a0a1-4e05-8260-020300e92286/registry-server/0.log" Feb 18 15:20:51 crc kubenswrapper[4896]: I0218 15:20:51.510507 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-77bdc966f6-nnk54_71d66b07-2d37-4b32-9d3e-6953073d0697/manager/0.log" Feb 18 15:20:51 crc kubenswrapper[4896]: I0218 15:20:51.568323 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-index-bgmmh_75da0d85-5bc2-41b4-8502-edd5b7dfd6ea/registry-server/0.log" Feb 18 15:20:51 crc kubenswrapper[4896]: I0218 15:20:51.602474 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qst58" Feb 18 15:20:51 crc kubenswrapper[4896]: I0218 15:20:51.602580 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qst58" Feb 18 15:20:51 crc kubenswrapper[4896]: I0218 15:20:51.649857 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qst58" Feb 18 15:20:51 crc kubenswrapper[4896]: I0218 15:20:51.764197 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-59746d8d7d-s27px_f1185065-2802-4e3e-a397-3170992e7b02/manager/0.log" Feb 18 15:20:51 crc kubenswrapper[4896]: I0218 15:20:51.800153 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-index-lqkgg_32c638c3-ec07-42c1-9473-8f0483c5de4c/registry-server/0.log" Feb 18 15:20:52 crc kubenswrapper[4896]: I0218 15:20:52.076940 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-779fc9694b-v9qdn_ddd714d9-66f9-4caa-8d0c-a6e0ff935af0/operator/0.log" Feb 18 15:20:52 crc kubenswrapper[4896]: I0218 15:20:52.147422 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-index-vtpwb_b56a5e00-18b9-42a9-9c39-a9a70fe7ab30/registry-server/0.log" Feb 18 15:20:52 crc kubenswrapper[4896]: I0218 15:20:52.320008 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-operator-index-ngrc4_5962f084-f27d-47b3-a9c1-b1389078c88b/registry-server/0.log" Feb 18 15:20:52 crc kubenswrapper[4896]: I0218 15:20:52.420758 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qst58" Feb 18 15:20:52 crc kubenswrapper[4896]: I0218 15:20:52.468530 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qst58"] Feb 18 15:20:54 crc kubenswrapper[4896]: I0218 15:20:54.305306 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bs6jw"] Feb 18 15:20:54 crc kubenswrapper[4896]: I0218 15:20:54.306764 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bs6jw" Feb 18 15:20:54 crc kubenswrapper[4896]: I0218 15:20:54.317699 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bs6jw"] Feb 18 15:20:54 crc kubenswrapper[4896]: I0218 15:20:54.324277 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de5ba3d1-e024-41c5-9ecb-295cd7f13171-catalog-content\") pod \"community-operators-bs6jw\" (UID: \"de5ba3d1-e024-41c5-9ecb-295cd7f13171\") " pod="openshift-marketplace/community-operators-bs6jw" Feb 18 15:20:54 crc kubenswrapper[4896]: I0218 15:20:54.324346 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de5ba3d1-e024-41c5-9ecb-295cd7f13171-utilities\") pod \"community-operators-bs6jw\" (UID: \"de5ba3d1-e024-41c5-9ecb-295cd7f13171\") " pod="openshift-marketplace/community-operators-bs6jw" Feb 18 15:20:54 crc kubenswrapper[4896]: I0218 15:20:54.324370 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbwzc\" (UniqueName: \"kubernetes.io/projected/de5ba3d1-e024-41c5-9ecb-295cd7f13171-kube-api-access-fbwzc\") pod \"community-operators-bs6jw\" (UID: \"de5ba3d1-e024-41c5-9ecb-295cd7f13171\") " pod="openshift-marketplace/community-operators-bs6jw" Feb 18 15:20:54 crc kubenswrapper[4896]: I0218 15:20:54.387997 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qst58" podUID="c51c69de-45a2-44ad-834d-47cb43e328b9" containerName="registry-server" containerID="cri-o://cd2cd75827c03c15067dd3f71cf1dd99348ca66610d41c21d341d5c7cd1f5a78" gracePeriod=2 Feb 18 15:20:54 crc kubenswrapper[4896]: I0218 15:20:54.426118 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de5ba3d1-e024-41c5-9ecb-295cd7f13171-catalog-content\") pod \"community-operators-bs6jw\" (UID: \"de5ba3d1-e024-41c5-9ecb-295cd7f13171\") " pod="openshift-marketplace/community-operators-bs6jw" Feb 18 15:20:54 crc kubenswrapper[4896]: I0218 15:20:54.426189 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de5ba3d1-e024-41c5-9ecb-295cd7f13171-utilities\") pod \"community-operators-bs6jw\" (UID: \"de5ba3d1-e024-41c5-9ecb-295cd7f13171\") " pod="openshift-marketplace/community-operators-bs6jw" Feb 18 15:20:54 crc kubenswrapper[4896]: I0218 15:20:54.426206 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbwzc\" (UniqueName: \"kubernetes.io/projected/de5ba3d1-e024-41c5-9ecb-295cd7f13171-kube-api-access-fbwzc\") pod \"community-operators-bs6jw\" (UID: \"de5ba3d1-e024-41c5-9ecb-295cd7f13171\") " pod="openshift-marketplace/community-operators-bs6jw" Feb 18 15:20:54 crc kubenswrapper[4896]: I0218 15:20:54.426614 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de5ba3d1-e024-41c5-9ecb-295cd7f13171-catalog-content\") pod \"community-operators-bs6jw\" (UID: \"de5ba3d1-e024-41c5-9ecb-295cd7f13171\") " pod="openshift-marketplace/community-operators-bs6jw" Feb 18 15:20:54 crc kubenswrapper[4896]: I0218 15:20:54.426698 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de5ba3d1-e024-41c5-9ecb-295cd7f13171-utilities\") pod \"community-operators-bs6jw\" (UID: \"de5ba3d1-e024-41c5-9ecb-295cd7f13171\") " pod="openshift-marketplace/community-operators-bs6jw" Feb 18 15:20:54 crc kubenswrapper[4896]: I0218 15:20:54.452490 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbwzc\" (UniqueName: \"kubernetes.io/projected/de5ba3d1-e024-41c5-9ecb-295cd7f13171-kube-api-access-fbwzc\") pod \"community-operators-bs6jw\" (UID: \"de5ba3d1-e024-41c5-9ecb-295cd7f13171\") " pod="openshift-marketplace/community-operators-bs6jw" Feb 18 15:20:54 crc kubenswrapper[4896]: I0218 15:20:54.682228 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bs6jw" Feb 18 15:20:55 crc kubenswrapper[4896]: I0218 15:20:55.164847 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bs6jw"] Feb 18 15:20:55 crc kubenswrapper[4896]: W0218 15:20:55.178047 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podde5ba3d1_e024_41c5_9ecb_295cd7f13171.slice/crio-513df927b78015121583288d8542491824c1c7d1b86196992d3eba0af11413d3 WatchSource:0}: Error finding container 513df927b78015121583288d8542491824c1c7d1b86196992d3eba0af11413d3: Status 404 returned error can't find the container with id 513df927b78015121583288d8542491824c1c7d1b86196992d3eba0af11413d3 Feb 18 15:20:55 crc kubenswrapper[4896]: I0218 15:20:55.394782 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bs6jw" event={"ID":"de5ba3d1-e024-41c5-9ecb-295cd7f13171","Type":"ContainerStarted","Data":"513df927b78015121583288d8542491824c1c7d1b86196992d3eba0af11413d3"} Feb 18 15:20:55 crc kubenswrapper[4896]: I0218 15:20:55.872403 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qst58" Feb 18 15:20:56 crc kubenswrapper[4896]: I0218 15:20:56.047703 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c51c69de-45a2-44ad-834d-47cb43e328b9-catalog-content\") pod \"c51c69de-45a2-44ad-834d-47cb43e328b9\" (UID: \"c51c69de-45a2-44ad-834d-47cb43e328b9\") " Feb 18 15:20:56 crc kubenswrapper[4896]: I0218 15:20:56.047768 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9qtf\" (UniqueName: \"kubernetes.io/projected/c51c69de-45a2-44ad-834d-47cb43e328b9-kube-api-access-z9qtf\") pod \"c51c69de-45a2-44ad-834d-47cb43e328b9\" (UID: \"c51c69de-45a2-44ad-834d-47cb43e328b9\") " Feb 18 15:20:56 crc kubenswrapper[4896]: I0218 15:20:56.047936 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c51c69de-45a2-44ad-834d-47cb43e328b9-utilities\") pod \"c51c69de-45a2-44ad-834d-47cb43e328b9\" (UID: \"c51c69de-45a2-44ad-834d-47cb43e328b9\") " Feb 18 15:20:56 crc kubenswrapper[4896]: I0218 15:20:56.049480 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c51c69de-45a2-44ad-834d-47cb43e328b9-utilities" (OuterVolumeSpecName: "utilities") pod "c51c69de-45a2-44ad-834d-47cb43e328b9" (UID: "c51c69de-45a2-44ad-834d-47cb43e328b9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:20:56 crc kubenswrapper[4896]: I0218 15:20:56.053570 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c51c69de-45a2-44ad-834d-47cb43e328b9-kube-api-access-z9qtf" (OuterVolumeSpecName: "kube-api-access-z9qtf") pod "c51c69de-45a2-44ad-834d-47cb43e328b9" (UID: "c51c69de-45a2-44ad-834d-47cb43e328b9"). InnerVolumeSpecName "kube-api-access-z9qtf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:20:56 crc kubenswrapper[4896]: I0218 15:20:56.149397 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9qtf\" (UniqueName: \"kubernetes.io/projected/c51c69de-45a2-44ad-834d-47cb43e328b9-kube-api-access-z9qtf\") on node \"crc\" DevicePath \"\"" Feb 18 15:20:56 crc kubenswrapper[4896]: I0218 15:20:56.149440 4896 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c51c69de-45a2-44ad-834d-47cb43e328b9-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 15:20:56 crc kubenswrapper[4896]: I0218 15:20:56.163156 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c51c69de-45a2-44ad-834d-47cb43e328b9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c51c69de-45a2-44ad-834d-47cb43e328b9" (UID: "c51c69de-45a2-44ad-834d-47cb43e328b9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:20:56 crc kubenswrapper[4896]: I0218 15:20:56.250669 4896 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c51c69de-45a2-44ad-834d-47cb43e328b9-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 15:20:56 crc kubenswrapper[4896]: I0218 15:20:56.402212 4896 generic.go:334] "Generic (PLEG): container finished" podID="de5ba3d1-e024-41c5-9ecb-295cd7f13171" containerID="de5ecc9e8a5a20cfb698e6b4b76e460a0a0948dc176d3f12d0b4df3b1df2604c" exitCode=0 Feb 18 15:20:56 crc kubenswrapper[4896]: I0218 15:20:56.402291 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bs6jw" event={"ID":"de5ba3d1-e024-41c5-9ecb-295cd7f13171","Type":"ContainerDied","Data":"de5ecc9e8a5a20cfb698e6b4b76e460a0a0948dc176d3f12d0b4df3b1df2604c"} Feb 18 15:20:56 crc kubenswrapper[4896]: I0218 15:20:56.405858 4896 generic.go:334] "Generic (PLEG): container finished" podID="c51c69de-45a2-44ad-834d-47cb43e328b9" containerID="cd2cd75827c03c15067dd3f71cf1dd99348ca66610d41c21d341d5c7cd1f5a78" exitCode=0 Feb 18 15:20:56 crc kubenswrapper[4896]: I0218 15:20:56.405972 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qst58" event={"ID":"c51c69de-45a2-44ad-834d-47cb43e328b9","Type":"ContainerDied","Data":"cd2cd75827c03c15067dd3f71cf1dd99348ca66610d41c21d341d5c7cd1f5a78"} Feb 18 15:20:56 crc kubenswrapper[4896]: I0218 15:20:56.406077 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qst58" event={"ID":"c51c69de-45a2-44ad-834d-47cb43e328b9","Type":"ContainerDied","Data":"027192e2a32397767c95ec863374f7d5d0df623d534bab229c229bd4ec0185bb"} Feb 18 15:20:56 crc kubenswrapper[4896]: I0218 15:20:56.406161 4896 scope.go:117] "RemoveContainer" containerID="cd2cd75827c03c15067dd3f71cf1dd99348ca66610d41c21d341d5c7cd1f5a78" Feb 18 15:20:56 crc kubenswrapper[4896]: I0218 15:20:56.406411 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qst58" Feb 18 15:20:56 crc kubenswrapper[4896]: I0218 15:20:56.428597 4896 scope.go:117] "RemoveContainer" containerID="b95e3b8fb24b92809bfecca896de913d5bd1df427c5014c6bcddc91d0890557d" Feb 18 15:20:56 crc kubenswrapper[4896]: I0218 15:20:56.464147 4896 scope.go:117] "RemoveContainer" containerID="c3fc429252aa3569b6127506375a1b2a9142768aa776246dc0b4c2fd72709e3c" Feb 18 15:20:56 crc kubenswrapper[4896]: I0218 15:20:56.492362 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qst58"] Feb 18 15:20:56 crc kubenswrapper[4896]: I0218 15:20:56.504790 4896 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qst58"] Feb 18 15:20:56 crc kubenswrapper[4896]: I0218 15:20:56.509381 4896 scope.go:117] "RemoveContainer" containerID="cd2cd75827c03c15067dd3f71cf1dd99348ca66610d41c21d341d5c7cd1f5a78" Feb 18 15:20:56 crc kubenswrapper[4896]: E0218 15:20:56.517388 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd2cd75827c03c15067dd3f71cf1dd99348ca66610d41c21d341d5c7cd1f5a78\": container with ID starting with cd2cd75827c03c15067dd3f71cf1dd99348ca66610d41c21d341d5c7cd1f5a78 not found: ID does not exist" containerID="cd2cd75827c03c15067dd3f71cf1dd99348ca66610d41c21d341d5c7cd1f5a78" Feb 18 15:20:56 crc kubenswrapper[4896]: I0218 15:20:56.517432 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd2cd75827c03c15067dd3f71cf1dd99348ca66610d41c21d341d5c7cd1f5a78"} err="failed to get container status \"cd2cd75827c03c15067dd3f71cf1dd99348ca66610d41c21d341d5c7cd1f5a78\": rpc error: code = NotFound desc = could not find container \"cd2cd75827c03c15067dd3f71cf1dd99348ca66610d41c21d341d5c7cd1f5a78\": container with ID starting with cd2cd75827c03c15067dd3f71cf1dd99348ca66610d41c21d341d5c7cd1f5a78 not found: ID does not exist" Feb 18 15:20:56 crc kubenswrapper[4896]: I0218 15:20:56.517457 4896 scope.go:117] "RemoveContainer" containerID="b95e3b8fb24b92809bfecca896de913d5bd1df427c5014c6bcddc91d0890557d" Feb 18 15:20:56 crc kubenswrapper[4896]: E0218 15:20:56.518075 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b95e3b8fb24b92809bfecca896de913d5bd1df427c5014c6bcddc91d0890557d\": container with ID starting with b95e3b8fb24b92809bfecca896de913d5bd1df427c5014c6bcddc91d0890557d not found: ID does not exist" containerID="b95e3b8fb24b92809bfecca896de913d5bd1df427c5014c6bcddc91d0890557d" Feb 18 15:20:56 crc kubenswrapper[4896]: I0218 15:20:56.518217 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b95e3b8fb24b92809bfecca896de913d5bd1df427c5014c6bcddc91d0890557d"} err="failed to get container status \"b95e3b8fb24b92809bfecca896de913d5bd1df427c5014c6bcddc91d0890557d\": rpc error: code = NotFound desc = could not find container \"b95e3b8fb24b92809bfecca896de913d5bd1df427c5014c6bcddc91d0890557d\": container with ID starting with b95e3b8fb24b92809bfecca896de913d5bd1df427c5014c6bcddc91d0890557d not found: ID does not exist" Feb 18 15:20:56 crc kubenswrapper[4896]: I0218 15:20:56.518344 4896 scope.go:117] "RemoveContainer" containerID="c3fc429252aa3569b6127506375a1b2a9142768aa776246dc0b4c2fd72709e3c" Feb 18 15:20:56 crc kubenswrapper[4896]: E0218 15:20:56.518620 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3fc429252aa3569b6127506375a1b2a9142768aa776246dc0b4c2fd72709e3c\": container with ID starting with c3fc429252aa3569b6127506375a1b2a9142768aa776246dc0b4c2fd72709e3c not found: ID does not exist" containerID="c3fc429252aa3569b6127506375a1b2a9142768aa776246dc0b4c2fd72709e3c" Feb 18 15:20:56 crc kubenswrapper[4896]: I0218 15:20:56.518646 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3fc429252aa3569b6127506375a1b2a9142768aa776246dc0b4c2fd72709e3c"} err="failed to get container status \"c3fc429252aa3569b6127506375a1b2a9142768aa776246dc0b4c2fd72709e3c\": rpc error: code = NotFound desc = could not find container \"c3fc429252aa3569b6127506375a1b2a9142768aa776246dc0b4c2fd72709e3c\": container with ID starting with c3fc429252aa3569b6127506375a1b2a9142768aa776246dc0b4c2fd72709e3c not found: ID does not exist" Feb 18 15:20:56 crc kubenswrapper[4896]: I0218 15:20:56.928418 4896 scope.go:117] "RemoveContainer" containerID="7cbf684d610fed53f6fdd655adb1877f0df69baa4d3a58b561f973625a82e83c" Feb 18 15:20:56 crc kubenswrapper[4896]: E0218 15:20:56.928778 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-24rq5_openshift-machine-config-operator(8503734a-8111-4fab-beda-1f0f3b59615b)\"" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" Feb 18 15:20:57 crc kubenswrapper[4896]: I0218 15:20:57.936685 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c51c69de-45a2-44ad-834d-47cb43e328b9" path="/var/lib/kubelet/pods/c51c69de-45a2-44ad-834d-47cb43e328b9/volumes" Feb 18 15:20:58 crc kubenswrapper[4896]: I0218 15:20:58.421071 4896 generic.go:334] "Generic (PLEG): container finished" podID="de5ba3d1-e024-41c5-9ecb-295cd7f13171" containerID="5b0ac780eb989483cd1c1dcad3b857b448c2fd047b2b29de7e82c87ce1563ac9" exitCode=0 Feb 18 15:20:58 crc kubenswrapper[4896]: I0218 15:20:58.421122 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bs6jw" event={"ID":"de5ba3d1-e024-41c5-9ecb-295cd7f13171","Type":"ContainerDied","Data":"5b0ac780eb989483cd1c1dcad3b857b448c2fd047b2b29de7e82c87ce1563ac9"} Feb 18 15:20:59 crc kubenswrapper[4896]: I0218 15:20:59.428805 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bs6jw" event={"ID":"de5ba3d1-e024-41c5-9ecb-295cd7f13171","Type":"ContainerStarted","Data":"3b9233df564f81d7a969ab26bc0cb3c962697897bf9105e23649a6437aa37910"} Feb 18 15:20:59 crc kubenswrapper[4896]: I0218 15:20:59.444974 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bs6jw" podStartSLOduration=3.004466294 podStartE2EDuration="5.444957552s" podCreationTimestamp="2026-02-18 15:20:54 +0000 UTC" firstStartedPulling="2026-02-18 15:20:56.404683381 +0000 UTC m=+1555.334017594" lastFinishedPulling="2026-02-18 15:20:58.845174639 +0000 UTC m=+1557.774508852" observedRunningTime="2026-02-18 15:20:59.442961025 +0000 UTC m=+1558.372295238" watchObservedRunningTime="2026-02-18 15:20:59.444957552 +0000 UTC m=+1558.374291765" Feb 18 15:21:04 crc kubenswrapper[4896]: I0218 15:21:04.682581 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bs6jw" Feb 18 15:21:04 crc kubenswrapper[4896]: I0218 15:21:04.683179 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bs6jw" Feb 18 15:21:04 crc kubenswrapper[4896]: I0218 15:21:04.727436 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bs6jw" Feb 18 15:21:05 crc kubenswrapper[4896]: I0218 15:21:05.506513 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bs6jw" Feb 18 15:21:05 crc kubenswrapper[4896]: I0218 15:21:05.559137 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bs6jw"] Feb 18 15:21:06 crc kubenswrapper[4896]: I0218 15:21:06.278334 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-jpg48_8c1a0565-e55b-410e-9b63-12ababdb69eb/control-plane-machine-set-operator/0.log" Feb 18 15:21:06 crc kubenswrapper[4896]: I0218 15:21:06.443754 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-22f85_42ba2ce9-d014-4090-b4b2-21a2ae8cea4c/machine-api-operator/0.log" Feb 18 15:21:06 crc kubenswrapper[4896]: I0218 15:21:06.459485 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-22f85_42ba2ce9-d014-4090-b4b2-21a2ae8cea4c/kube-rbac-proxy/0.log" Feb 18 15:21:07 crc kubenswrapper[4896]: I0218 15:21:07.474914 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bs6jw" podUID="de5ba3d1-e024-41c5-9ecb-295cd7f13171" containerName="registry-server" containerID="cri-o://3b9233df564f81d7a969ab26bc0cb3c962697897bf9105e23649a6437aa37910" gracePeriod=2 Feb 18 15:21:07 crc kubenswrapper[4896]: I0218 15:21:07.901241 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bs6jw" Feb 18 15:21:08 crc kubenswrapper[4896]: I0218 15:21:08.026041 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de5ba3d1-e024-41c5-9ecb-295cd7f13171-catalog-content\") pod \"de5ba3d1-e024-41c5-9ecb-295cd7f13171\" (UID: \"de5ba3d1-e024-41c5-9ecb-295cd7f13171\") " Feb 18 15:21:08 crc kubenswrapper[4896]: I0218 15:21:08.026140 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de5ba3d1-e024-41c5-9ecb-295cd7f13171-utilities\") pod \"de5ba3d1-e024-41c5-9ecb-295cd7f13171\" (UID: \"de5ba3d1-e024-41c5-9ecb-295cd7f13171\") " Feb 18 15:21:08 crc kubenswrapper[4896]: I0218 15:21:08.026170 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbwzc\" (UniqueName: \"kubernetes.io/projected/de5ba3d1-e024-41c5-9ecb-295cd7f13171-kube-api-access-fbwzc\") pod \"de5ba3d1-e024-41c5-9ecb-295cd7f13171\" (UID: \"de5ba3d1-e024-41c5-9ecb-295cd7f13171\") " Feb 18 15:21:08 crc kubenswrapper[4896]: I0218 15:21:08.027171 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de5ba3d1-e024-41c5-9ecb-295cd7f13171-utilities" (OuterVolumeSpecName: "utilities") pod "de5ba3d1-e024-41c5-9ecb-295cd7f13171" (UID: "de5ba3d1-e024-41c5-9ecb-295cd7f13171"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:21:08 crc kubenswrapper[4896]: I0218 15:21:08.035403 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de5ba3d1-e024-41c5-9ecb-295cd7f13171-kube-api-access-fbwzc" (OuterVolumeSpecName: "kube-api-access-fbwzc") pod "de5ba3d1-e024-41c5-9ecb-295cd7f13171" (UID: "de5ba3d1-e024-41c5-9ecb-295cd7f13171"). InnerVolumeSpecName "kube-api-access-fbwzc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:21:08 crc kubenswrapper[4896]: I0218 15:21:08.075507 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de5ba3d1-e024-41c5-9ecb-295cd7f13171-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "de5ba3d1-e024-41c5-9ecb-295cd7f13171" (UID: "de5ba3d1-e024-41c5-9ecb-295cd7f13171"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:21:08 crc kubenswrapper[4896]: I0218 15:21:08.127991 4896 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de5ba3d1-e024-41c5-9ecb-295cd7f13171-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 15:21:08 crc kubenswrapper[4896]: I0218 15:21:08.128026 4896 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de5ba3d1-e024-41c5-9ecb-295cd7f13171-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 15:21:08 crc kubenswrapper[4896]: I0218 15:21:08.128037 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbwzc\" (UniqueName: \"kubernetes.io/projected/de5ba3d1-e024-41c5-9ecb-295cd7f13171-kube-api-access-fbwzc\") on node \"crc\" DevicePath \"\"" Feb 18 15:21:08 crc kubenswrapper[4896]: I0218 15:21:08.482759 4896 generic.go:334] "Generic (PLEG): container finished" podID="de5ba3d1-e024-41c5-9ecb-295cd7f13171" containerID="3b9233df564f81d7a969ab26bc0cb3c962697897bf9105e23649a6437aa37910" exitCode=0 Feb 18 15:21:08 crc kubenswrapper[4896]: I0218 15:21:08.482800 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bs6jw" event={"ID":"de5ba3d1-e024-41c5-9ecb-295cd7f13171","Type":"ContainerDied","Data":"3b9233df564f81d7a969ab26bc0cb3c962697897bf9105e23649a6437aa37910"} Feb 18 15:21:08 crc kubenswrapper[4896]: I0218 15:21:08.482853 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bs6jw" Feb 18 15:21:08 crc kubenswrapper[4896]: I0218 15:21:08.482875 4896 scope.go:117] "RemoveContainer" containerID="3b9233df564f81d7a969ab26bc0cb3c962697897bf9105e23649a6437aa37910" Feb 18 15:21:08 crc kubenswrapper[4896]: I0218 15:21:08.482858 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bs6jw" event={"ID":"de5ba3d1-e024-41c5-9ecb-295cd7f13171","Type":"ContainerDied","Data":"513df927b78015121583288d8542491824c1c7d1b86196992d3eba0af11413d3"} Feb 18 15:21:08 crc kubenswrapper[4896]: I0218 15:21:08.511085 4896 scope.go:117] "RemoveContainer" containerID="5b0ac780eb989483cd1c1dcad3b857b448c2fd047b2b29de7e82c87ce1563ac9" Feb 18 15:21:08 crc kubenswrapper[4896]: I0218 15:21:08.536081 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bs6jw"] Feb 18 15:21:08 crc kubenswrapper[4896]: I0218 15:21:08.541408 4896 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bs6jw"] Feb 18 15:21:08 crc kubenswrapper[4896]: I0218 15:21:08.561779 4896 scope.go:117] "RemoveContainer" containerID="de5ecc9e8a5a20cfb698e6b4b76e460a0a0948dc176d3f12d0b4df3b1df2604c" Feb 18 15:21:08 crc kubenswrapper[4896]: I0218 15:21:08.597329 4896 scope.go:117] "RemoveContainer" containerID="3b9233df564f81d7a969ab26bc0cb3c962697897bf9105e23649a6437aa37910" Feb 18 15:21:08 crc kubenswrapper[4896]: E0218 15:21:08.598798 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b9233df564f81d7a969ab26bc0cb3c962697897bf9105e23649a6437aa37910\": container with ID starting with 3b9233df564f81d7a969ab26bc0cb3c962697897bf9105e23649a6437aa37910 not found: ID does not exist" containerID="3b9233df564f81d7a969ab26bc0cb3c962697897bf9105e23649a6437aa37910" Feb 18 15:21:08 crc kubenswrapper[4896]: I0218 15:21:08.598839 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b9233df564f81d7a969ab26bc0cb3c962697897bf9105e23649a6437aa37910"} err="failed to get container status \"3b9233df564f81d7a969ab26bc0cb3c962697897bf9105e23649a6437aa37910\": rpc error: code = NotFound desc = could not find container \"3b9233df564f81d7a969ab26bc0cb3c962697897bf9105e23649a6437aa37910\": container with ID starting with 3b9233df564f81d7a969ab26bc0cb3c962697897bf9105e23649a6437aa37910 not found: ID does not exist" Feb 18 15:21:08 crc kubenswrapper[4896]: I0218 15:21:08.598866 4896 scope.go:117] "RemoveContainer" containerID="5b0ac780eb989483cd1c1dcad3b857b448c2fd047b2b29de7e82c87ce1563ac9" Feb 18 15:21:08 crc kubenswrapper[4896]: E0218 15:21:08.603986 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b0ac780eb989483cd1c1dcad3b857b448c2fd047b2b29de7e82c87ce1563ac9\": container with ID starting with 5b0ac780eb989483cd1c1dcad3b857b448c2fd047b2b29de7e82c87ce1563ac9 not found: ID does not exist" containerID="5b0ac780eb989483cd1c1dcad3b857b448c2fd047b2b29de7e82c87ce1563ac9" Feb 18 15:21:08 crc kubenswrapper[4896]: I0218 15:21:08.604039 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b0ac780eb989483cd1c1dcad3b857b448c2fd047b2b29de7e82c87ce1563ac9"} err="failed to get container status \"5b0ac780eb989483cd1c1dcad3b857b448c2fd047b2b29de7e82c87ce1563ac9\": rpc error: code = NotFound desc = could not find container \"5b0ac780eb989483cd1c1dcad3b857b448c2fd047b2b29de7e82c87ce1563ac9\": container with ID starting with 5b0ac780eb989483cd1c1dcad3b857b448c2fd047b2b29de7e82c87ce1563ac9 not found: ID does not exist" Feb 18 15:21:08 crc kubenswrapper[4896]: I0218 15:21:08.604067 4896 scope.go:117] "RemoveContainer" containerID="de5ecc9e8a5a20cfb698e6b4b76e460a0a0948dc176d3f12d0b4df3b1df2604c" Feb 18 15:21:08 crc kubenswrapper[4896]: E0218 15:21:08.606764 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de5ecc9e8a5a20cfb698e6b4b76e460a0a0948dc176d3f12d0b4df3b1df2604c\": container with ID starting with de5ecc9e8a5a20cfb698e6b4b76e460a0a0948dc176d3f12d0b4df3b1df2604c not found: ID does not exist" containerID="de5ecc9e8a5a20cfb698e6b4b76e460a0a0948dc176d3f12d0b4df3b1df2604c" Feb 18 15:21:08 crc kubenswrapper[4896]: I0218 15:21:08.606840 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de5ecc9e8a5a20cfb698e6b4b76e460a0a0948dc176d3f12d0b4df3b1df2604c"} err="failed to get container status \"de5ecc9e8a5a20cfb698e6b4b76e460a0a0948dc176d3f12d0b4df3b1df2604c\": rpc error: code = NotFound desc = could not find container \"de5ecc9e8a5a20cfb698e6b4b76e460a0a0948dc176d3f12d0b4df3b1df2604c\": container with ID starting with de5ecc9e8a5a20cfb698e6b4b76e460a0a0948dc176d3f12d0b4df3b1df2604c not found: ID does not exist" Feb 18 15:21:09 crc kubenswrapper[4896]: I0218 15:21:09.927877 4896 scope.go:117] "RemoveContainer" containerID="7cbf684d610fed53f6fdd655adb1877f0df69baa4d3a58b561f973625a82e83c" Feb 18 15:21:09 crc kubenswrapper[4896]: E0218 15:21:09.928508 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-24rq5_openshift-machine-config-operator(8503734a-8111-4fab-beda-1f0f3b59615b)\"" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" Feb 18 15:21:09 crc kubenswrapper[4896]: I0218 15:21:09.934819 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de5ba3d1-e024-41c5-9ecb-295cd7f13171" path="/var/lib/kubelet/pods/de5ba3d1-e024-41c5-9ecb-295cd7f13171/volumes" Feb 18 15:21:12 crc kubenswrapper[4896]: I0218 15:21:12.039456 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["test-operator-kuttl-tests/keystone-20ea-account-create-update-9qws6"] Feb 18 15:21:12 crc kubenswrapper[4896]: I0218 15:21:12.051318 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["test-operator-kuttl-tests/keystone-db-create-h8qzx"] Feb 18 15:21:12 crc kubenswrapper[4896]: I0218 15:21:12.059053 4896 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["test-operator-kuttl-tests/keystone-20ea-account-create-update-9qws6"] Feb 18 15:21:12 crc kubenswrapper[4896]: I0218 15:21:12.065490 4896 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["test-operator-kuttl-tests/keystone-db-create-h8qzx"] Feb 18 15:21:13 crc kubenswrapper[4896]: I0218 15:21:13.935064 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65ed11b1-ec28-4b2a-b755-6d324c33426c" path="/var/lib/kubelet/pods/65ed11b1-ec28-4b2a-b755-6d324c33426c/volumes" Feb 18 15:21:13 crc kubenswrapper[4896]: I0218 15:21:13.935928 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc590030-d392-4cd4-96f5-d28039c0faa6" path="/var/lib/kubelet/pods/dc590030-d392-4cd4-96f5-d28039c0faa6/volumes" Feb 18 15:21:22 crc kubenswrapper[4896]: I0218 15:21:22.927665 4896 scope.go:117] "RemoveContainer" containerID="7cbf684d610fed53f6fdd655adb1877f0df69baa4d3a58b561f973625a82e83c" Feb 18 15:21:22 crc kubenswrapper[4896]: E0218 15:21:22.928340 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-24rq5_openshift-machine-config-operator(8503734a-8111-4fab-beda-1f0f3b59615b)\"" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" Feb 18 15:21:28 crc kubenswrapper[4896]: I0218 15:21:28.025122 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["test-operator-kuttl-tests/keystone-db-sync-nf7cx"] Feb 18 15:21:28 crc kubenswrapper[4896]: I0218 15:21:28.033693 4896 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["test-operator-kuttl-tests/keystone-db-sync-nf7cx"] Feb 18 15:21:29 crc kubenswrapper[4896]: I0218 15:21:29.934696 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74496765-f69d-4a41-add9-5afe29af7e20" path="/var/lib/kubelet/pods/74496765-f69d-4a41-add9-5afe29af7e20/volumes" Feb 18 15:21:34 crc kubenswrapper[4896]: I0218 15:21:34.696044 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-zqgw4_7d817dfd-f711-48b3-8630-83d4b1a784de/kube-rbac-proxy/0.log" Feb 18 15:21:34 crc kubenswrapper[4896]: I0218 15:21:34.757854 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-69bbfbf88f-zqgw4_7d817dfd-f711-48b3-8630-83d4b1a784de/controller/0.log" Feb 18 15:21:34 crc kubenswrapper[4896]: I0218 15:21:34.916603 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lnrk2_29e01ca8-ce78-4769-a5d5-453834a27bb3/cp-frr-files/0.log" Feb 18 15:21:35 crc kubenswrapper[4896]: I0218 15:21:35.022142 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["test-operator-kuttl-tests/keystone-bootstrap-bmd8m"] Feb 18 15:21:35 crc kubenswrapper[4896]: I0218 15:21:35.026891 4896 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["test-operator-kuttl-tests/keystone-bootstrap-bmd8m"] Feb 18 15:21:35 crc kubenswrapper[4896]: I0218 15:21:35.055525 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lnrk2_29e01ca8-ce78-4769-a5d5-453834a27bb3/cp-frr-files/0.log" Feb 18 15:21:35 crc kubenswrapper[4896]: I0218 15:21:35.084594 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lnrk2_29e01ca8-ce78-4769-a5d5-453834a27bb3/cp-reloader/0.log" Feb 18 15:21:35 crc kubenswrapper[4896]: I0218 15:21:35.116761 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lnrk2_29e01ca8-ce78-4769-a5d5-453834a27bb3/cp-metrics/0.log" Feb 18 15:21:35 crc kubenswrapper[4896]: I0218 15:21:35.141708 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lnrk2_29e01ca8-ce78-4769-a5d5-453834a27bb3/cp-reloader/0.log" Feb 18 15:21:35 crc kubenswrapper[4896]: I0218 15:21:35.276288 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lnrk2_29e01ca8-ce78-4769-a5d5-453834a27bb3/cp-frr-files/0.log" Feb 18 15:21:35 crc kubenswrapper[4896]: I0218 15:21:35.308503 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lnrk2_29e01ca8-ce78-4769-a5d5-453834a27bb3/cp-reloader/0.log" Feb 18 15:21:35 crc kubenswrapper[4896]: I0218 15:21:35.350206 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lnrk2_29e01ca8-ce78-4769-a5d5-453834a27bb3/cp-metrics/0.log" Feb 18 15:21:35 crc kubenswrapper[4896]: I0218 15:21:35.369935 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lnrk2_29e01ca8-ce78-4769-a5d5-453834a27bb3/cp-metrics/0.log" Feb 18 15:21:35 crc kubenswrapper[4896]: I0218 15:21:35.542713 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lnrk2_29e01ca8-ce78-4769-a5d5-453834a27bb3/cp-frr-files/0.log" Feb 18 15:21:35 crc kubenswrapper[4896]: I0218 15:21:35.543136 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lnrk2_29e01ca8-ce78-4769-a5d5-453834a27bb3/cp-reloader/0.log" Feb 18 15:21:35 crc kubenswrapper[4896]: I0218 15:21:35.550217 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lnrk2_29e01ca8-ce78-4769-a5d5-453834a27bb3/cp-metrics/0.log" Feb 18 15:21:35 crc kubenswrapper[4896]: I0218 15:21:35.552835 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lnrk2_29e01ca8-ce78-4769-a5d5-453834a27bb3/controller/0.log" Feb 18 15:21:35 crc kubenswrapper[4896]: I0218 15:21:35.728444 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lnrk2_29e01ca8-ce78-4769-a5d5-453834a27bb3/kube-rbac-proxy-frr/0.log" Feb 18 15:21:35 crc kubenswrapper[4896]: I0218 15:21:35.744269 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lnrk2_29e01ca8-ce78-4769-a5d5-453834a27bb3/kube-rbac-proxy/0.log" Feb 18 15:21:35 crc kubenswrapper[4896]: I0218 15:21:35.747796 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lnrk2_29e01ca8-ce78-4769-a5d5-453834a27bb3/frr-metrics/0.log" Feb 18 15:21:35 crc kubenswrapper[4896]: I0218 15:21:35.909473 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lnrk2_29e01ca8-ce78-4769-a5d5-453834a27bb3/frr/0.log" Feb 18 15:21:35 crc kubenswrapper[4896]: I0218 15:21:35.938683 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d53c04c-f334-4252-91e3-8feb31200ee7" path="/var/lib/kubelet/pods/2d53c04c-f334-4252-91e3-8feb31200ee7/volumes" Feb 18 15:21:35 crc kubenswrapper[4896]: I0218 15:21:35.987117 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-78b44bf5bb-jxgkk_c2a8e7f2-bf46-4f7d-a7e0-c1373f5c2136/frr-k8s-webhook-server/0.log" Feb 18 15:21:35 crc kubenswrapper[4896]: I0218 15:21:35.994168 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lnrk2_29e01ca8-ce78-4769-a5d5-453834a27bb3/reloader/0.log" Feb 18 15:21:36 crc kubenswrapper[4896]: I0218 15:21:36.132688 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-756bb4b59-vd48g_60827093-1be2-4ca9-b9d3-3caaf37ebe83/manager/0.log" Feb 18 15:21:36 crc kubenswrapper[4896]: I0218 15:21:36.143525 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-d66b4dc77-fqp6g_23fd11e5-8c0f-434f-a334-01bf16e764e8/webhook-server/0.log" Feb 18 15:21:36 crc kubenswrapper[4896]: I0218 15:21:36.323223 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qb82t_d2436da1-6ba8-4028-9012-1160a473af76/kube-rbac-proxy/0.log" Feb 18 15:21:36 crc kubenswrapper[4896]: I0218 15:21:36.375291 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-qb82t_d2436da1-6ba8-4028-9012-1160a473af76/speaker/0.log" Feb 18 15:21:36 crc kubenswrapper[4896]: I0218 15:21:36.928568 4896 scope.go:117] "RemoveContainer" containerID="7cbf684d610fed53f6fdd655adb1877f0df69baa4d3a58b561f973625a82e83c" Feb 18 15:21:36 crc kubenswrapper[4896]: E0218 15:21:36.929054 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-24rq5_openshift-machine-config-operator(8503734a-8111-4fab-beda-1f0f3b59615b)\"" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" Feb 18 15:21:44 crc kubenswrapper[4896]: I0218 15:21:44.886096 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wlxrx"] Feb 18 15:21:44 crc kubenswrapper[4896]: E0218 15:21:44.886936 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c51c69de-45a2-44ad-834d-47cb43e328b9" containerName="extract-content" Feb 18 15:21:44 crc kubenswrapper[4896]: I0218 15:21:44.886950 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="c51c69de-45a2-44ad-834d-47cb43e328b9" containerName="extract-content" Feb 18 15:21:44 crc kubenswrapper[4896]: E0218 15:21:44.886968 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c51c69de-45a2-44ad-834d-47cb43e328b9" containerName="extract-utilities" Feb 18 15:21:44 crc kubenswrapper[4896]: I0218 15:21:44.886978 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="c51c69de-45a2-44ad-834d-47cb43e328b9" containerName="extract-utilities" Feb 18 15:21:44 crc kubenswrapper[4896]: E0218 15:21:44.886999 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de5ba3d1-e024-41c5-9ecb-295cd7f13171" containerName="extract-content" Feb 18 15:21:44 crc kubenswrapper[4896]: I0218 15:21:44.887007 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="de5ba3d1-e024-41c5-9ecb-295cd7f13171" containerName="extract-content" Feb 18 15:21:44 crc kubenswrapper[4896]: E0218 15:21:44.887023 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de5ba3d1-e024-41c5-9ecb-295cd7f13171" containerName="extract-utilities" Feb 18 15:21:44 crc kubenswrapper[4896]: I0218 15:21:44.887031 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="de5ba3d1-e024-41c5-9ecb-295cd7f13171" containerName="extract-utilities" Feb 18 15:21:44 crc kubenswrapper[4896]: E0218 15:21:44.887046 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c51c69de-45a2-44ad-834d-47cb43e328b9" containerName="registry-server" Feb 18 15:21:44 crc kubenswrapper[4896]: I0218 15:21:44.887054 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="c51c69de-45a2-44ad-834d-47cb43e328b9" containerName="registry-server" Feb 18 15:21:44 crc kubenswrapper[4896]: E0218 15:21:44.887066 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de5ba3d1-e024-41c5-9ecb-295cd7f13171" containerName="registry-server" Feb 18 15:21:44 crc kubenswrapper[4896]: I0218 15:21:44.887074 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="de5ba3d1-e024-41c5-9ecb-295cd7f13171" containerName="registry-server" Feb 18 15:21:44 crc kubenswrapper[4896]: I0218 15:21:44.887213 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="c51c69de-45a2-44ad-834d-47cb43e328b9" containerName="registry-server" Feb 18 15:21:44 crc kubenswrapper[4896]: I0218 15:21:44.887229 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="de5ba3d1-e024-41c5-9ecb-295cd7f13171" containerName="registry-server" Feb 18 15:21:44 crc kubenswrapper[4896]: I0218 15:21:44.888360 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wlxrx" Feb 18 15:21:44 crc kubenswrapper[4896]: I0218 15:21:44.908655 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wlxrx"] Feb 18 15:21:44 crc kubenswrapper[4896]: I0218 15:21:44.927082 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd46p\" (UniqueName: \"kubernetes.io/projected/d46a4767-2e61-4843-806b-4ac0c0d14d4e-kube-api-access-qd46p\") pod \"certified-operators-wlxrx\" (UID: \"d46a4767-2e61-4843-806b-4ac0c0d14d4e\") " pod="openshift-marketplace/certified-operators-wlxrx" Feb 18 15:21:44 crc kubenswrapper[4896]: I0218 15:21:44.927145 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d46a4767-2e61-4843-806b-4ac0c0d14d4e-utilities\") pod \"certified-operators-wlxrx\" (UID: \"d46a4767-2e61-4843-806b-4ac0c0d14d4e\") " pod="openshift-marketplace/certified-operators-wlxrx" Feb 18 15:21:44 crc kubenswrapper[4896]: I0218 15:21:44.927442 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d46a4767-2e61-4843-806b-4ac0c0d14d4e-catalog-content\") pod \"certified-operators-wlxrx\" (UID: \"d46a4767-2e61-4843-806b-4ac0c0d14d4e\") " pod="openshift-marketplace/certified-operators-wlxrx" Feb 18 15:21:45 crc kubenswrapper[4896]: I0218 15:21:45.028349 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d46a4767-2e61-4843-806b-4ac0c0d14d4e-catalog-content\") pod \"certified-operators-wlxrx\" (UID: \"d46a4767-2e61-4843-806b-4ac0c0d14d4e\") " pod="openshift-marketplace/certified-operators-wlxrx" Feb 18 15:21:45 crc kubenswrapper[4896]: I0218 15:21:45.028448 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qd46p\" (UniqueName: \"kubernetes.io/projected/d46a4767-2e61-4843-806b-4ac0c0d14d4e-kube-api-access-qd46p\") pod \"certified-operators-wlxrx\" (UID: \"d46a4767-2e61-4843-806b-4ac0c0d14d4e\") " pod="openshift-marketplace/certified-operators-wlxrx" Feb 18 15:21:45 crc kubenswrapper[4896]: I0218 15:21:45.028492 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d46a4767-2e61-4843-806b-4ac0c0d14d4e-utilities\") pod \"certified-operators-wlxrx\" (UID: \"d46a4767-2e61-4843-806b-4ac0c0d14d4e\") " pod="openshift-marketplace/certified-operators-wlxrx" Feb 18 15:21:45 crc kubenswrapper[4896]: I0218 15:21:45.028880 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d46a4767-2e61-4843-806b-4ac0c0d14d4e-catalog-content\") pod \"certified-operators-wlxrx\" (UID: \"d46a4767-2e61-4843-806b-4ac0c0d14d4e\") " pod="openshift-marketplace/certified-operators-wlxrx" Feb 18 15:21:45 crc kubenswrapper[4896]: I0218 15:21:45.028911 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d46a4767-2e61-4843-806b-4ac0c0d14d4e-utilities\") pod \"certified-operators-wlxrx\" (UID: \"d46a4767-2e61-4843-806b-4ac0c0d14d4e\") " pod="openshift-marketplace/certified-operators-wlxrx" Feb 18 15:21:45 crc kubenswrapper[4896]: I0218 15:21:45.048226 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qd46p\" (UniqueName: \"kubernetes.io/projected/d46a4767-2e61-4843-806b-4ac0c0d14d4e-kube-api-access-qd46p\") pod \"certified-operators-wlxrx\" (UID: \"d46a4767-2e61-4843-806b-4ac0c0d14d4e\") " pod="openshift-marketplace/certified-operators-wlxrx" Feb 18 15:21:45 crc kubenswrapper[4896]: I0218 15:21:45.205448 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wlxrx" Feb 18 15:21:45 crc kubenswrapper[4896]: I0218 15:21:45.679749 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wlxrx"] Feb 18 15:21:45 crc kubenswrapper[4896]: I0218 15:21:45.698266 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wlxrx" event={"ID":"d46a4767-2e61-4843-806b-4ac0c0d14d4e","Type":"ContainerStarted","Data":"6fca9e5f75c0abe0ed9dbd9f627ad8ce47bcc35a807ca2d4a824b5ba8786f1f0"} Feb 18 15:21:46 crc kubenswrapper[4896]: I0218 15:21:46.704982 4896 generic.go:334] "Generic (PLEG): container finished" podID="d46a4767-2e61-4843-806b-4ac0c0d14d4e" containerID="71c0925514e8db0fccce9b34a073fb43c41fac470f44f964e2af36775e772252" exitCode=0 Feb 18 15:21:46 crc kubenswrapper[4896]: I0218 15:21:46.705090 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wlxrx" event={"ID":"d46a4767-2e61-4843-806b-4ac0c0d14d4e","Type":"ContainerDied","Data":"71c0925514e8db0fccce9b34a073fb43c41fac470f44f964e2af36775e772252"} Feb 18 15:21:47 crc kubenswrapper[4896]: I0218 15:21:47.712532 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wlxrx" event={"ID":"d46a4767-2e61-4843-806b-4ac0c0d14d4e","Type":"ContainerStarted","Data":"e66ecb62c4321d2e4826c3d0ce3462595362e06bdb64068398a857fb137c5c1f"} Feb 18 15:21:48 crc kubenswrapper[4896]: I0218 15:21:48.720719 4896 generic.go:334] "Generic (PLEG): container finished" podID="d46a4767-2e61-4843-806b-4ac0c0d14d4e" containerID="e66ecb62c4321d2e4826c3d0ce3462595362e06bdb64068398a857fb137c5c1f" exitCode=0 Feb 18 15:21:48 crc kubenswrapper[4896]: I0218 15:21:48.720800 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wlxrx" event={"ID":"d46a4767-2e61-4843-806b-4ac0c0d14d4e","Type":"ContainerDied","Data":"e66ecb62c4321d2e4826c3d0ce3462595362e06bdb64068398a857fb137c5c1f"} Feb 18 15:21:48 crc kubenswrapper[4896]: I0218 15:21:48.928546 4896 scope.go:117] "RemoveContainer" containerID="7cbf684d610fed53f6fdd655adb1877f0df69baa4d3a58b561f973625a82e83c" Feb 18 15:21:48 crc kubenswrapper[4896]: E0218 15:21:48.928810 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-24rq5_openshift-machine-config-operator(8503734a-8111-4fab-beda-1f0f3b59615b)\"" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" Feb 18 15:21:49 crc kubenswrapper[4896]: I0218 15:21:49.269752 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/test-operator-kuttl-tests_keystone-7454c655dc-bjql4_7b99a0f0-1581-40ba-b1a9-6311ac1be415/keystone-api/0.log" Feb 18 15:21:49 crc kubenswrapper[4896]: I0218 15:21:49.401083 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/test-operator-kuttl-tests_memcached-0_7dfbd600-c900-4217-a46a-cb07e68791e2/memcached/0.log" Feb 18 15:21:49 crc kubenswrapper[4896]: I0218 15:21:49.533124 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/test-operator-kuttl-tests_openstack-galera-0_1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed/mysql-bootstrap/0.log" Feb 18 15:21:49 crc kubenswrapper[4896]: I0218 15:21:49.728768 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wlxrx" event={"ID":"d46a4767-2e61-4843-806b-4ac0c0d14d4e","Type":"ContainerStarted","Data":"69aa7283c33ce19724aae11be6050d24b29018149303c0d40252027b6165c3ed"} Feb 18 15:21:49 crc kubenswrapper[4896]: I0218 15:21:49.729521 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/test-operator-kuttl-tests_openstack-galera-0_1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed/galera/0.log" Feb 18 15:21:49 crc kubenswrapper[4896]: I0218 15:21:49.779359 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wlxrx" podStartSLOduration=3.351763691 podStartE2EDuration="5.779339556s" podCreationTimestamp="2026-02-18 15:21:44 +0000 UTC" firstStartedPulling="2026-02-18 15:21:46.706627963 +0000 UTC m=+1605.635962176" lastFinishedPulling="2026-02-18 15:21:49.134203828 +0000 UTC m=+1608.063538041" observedRunningTime="2026-02-18 15:21:49.774351105 +0000 UTC m=+1608.703685318" watchObservedRunningTime="2026-02-18 15:21:49.779339556 +0000 UTC m=+1608.708673769" Feb 18 15:21:49 crc kubenswrapper[4896]: I0218 15:21:49.833376 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/test-operator-kuttl-tests_openstack-galera-0_1fb2a5b0-19cb-4d59-8d9d-13ffb1c499ed/mysql-bootstrap/0.log" Feb 18 15:21:49 crc kubenswrapper[4896]: I0218 15:21:49.857109 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/test-operator-kuttl-tests_openstack-galera-1_1a1927a6-afa4-41c7-95ad-22d7bba44d38/mysql-bootstrap/0.log" Feb 18 15:21:50 crc kubenswrapper[4896]: I0218 15:21:50.021628 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/test-operator-kuttl-tests_openstack-galera-1_1a1927a6-afa4-41c7-95ad-22d7bba44d38/galera/0.log" Feb 18 15:21:50 crc kubenswrapper[4896]: I0218 15:21:50.035872 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/test-operator-kuttl-tests_openstack-galera-1_1a1927a6-afa4-41c7-95ad-22d7bba44d38/mysql-bootstrap/0.log" Feb 18 15:21:50 crc kubenswrapper[4896]: I0218 15:21:50.142721 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/test-operator-kuttl-tests_openstack-galera-2_0e13f33c-acb4-421a-be84-4757ff7c3b74/mysql-bootstrap/0.log" Feb 18 15:21:50 crc kubenswrapper[4896]: I0218 15:21:50.310982 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/test-operator-kuttl-tests_openstack-galera-2_0e13f33c-acb4-421a-be84-4757ff7c3b74/mysql-bootstrap/0.log" Feb 18 15:21:50 crc kubenswrapper[4896]: I0218 15:21:50.338811 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/test-operator-kuttl-tests_openstack-galera-2_0e13f33c-acb4-421a-be84-4757ff7c3b74/galera/0.log" Feb 18 15:21:50 crc kubenswrapper[4896]: I0218 15:21:50.437616 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/test-operator-kuttl-tests_rabbitmq-server-0_c5ce908c-bcd7-456e-9ab2-abe730180641/setup-container/0.log" Feb 18 15:21:50 crc kubenswrapper[4896]: I0218 15:21:50.579973 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/test-operator-kuttl-tests_rabbitmq-server-0_c5ce908c-bcd7-456e-9ab2-abe730180641/setup-container/0.log" Feb 18 15:21:50 crc kubenswrapper[4896]: I0218 15:21:50.587227 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/test-operator-kuttl-tests_rabbitmq-server-0_c5ce908c-bcd7-456e-9ab2-abe730180641/rabbitmq/0.log" Feb 18 15:21:55 crc kubenswrapper[4896]: I0218 15:21:55.206322 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wlxrx" Feb 18 15:21:55 crc kubenswrapper[4896]: I0218 15:21:55.206818 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wlxrx" Feb 18 15:21:55 crc kubenswrapper[4896]: I0218 15:21:55.250640 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wlxrx" Feb 18 15:21:55 crc kubenswrapper[4896]: I0218 15:21:55.807037 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wlxrx" Feb 18 15:21:55 crc kubenswrapper[4896]: I0218 15:21:55.861584 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wlxrx"] Feb 18 15:21:57 crc kubenswrapper[4896]: I0218 15:21:57.782330 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wlxrx" podUID="d46a4767-2e61-4843-806b-4ac0c0d14d4e" containerName="registry-server" containerID="cri-o://69aa7283c33ce19724aae11be6050d24b29018149303c0d40252027b6165c3ed" gracePeriod=2 Feb 18 15:21:58 crc kubenswrapper[4896]: I0218 15:21:58.135449 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wlxrx" Feb 18 15:21:58 crc kubenswrapper[4896]: I0218 15:21:58.248820 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qd46p\" (UniqueName: \"kubernetes.io/projected/d46a4767-2e61-4843-806b-4ac0c0d14d4e-kube-api-access-qd46p\") pod \"d46a4767-2e61-4843-806b-4ac0c0d14d4e\" (UID: \"d46a4767-2e61-4843-806b-4ac0c0d14d4e\") " Feb 18 15:21:58 crc kubenswrapper[4896]: I0218 15:21:58.248936 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d46a4767-2e61-4843-806b-4ac0c0d14d4e-utilities\") pod \"d46a4767-2e61-4843-806b-4ac0c0d14d4e\" (UID: \"d46a4767-2e61-4843-806b-4ac0c0d14d4e\") " Feb 18 15:21:58 crc kubenswrapper[4896]: I0218 15:21:58.249007 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d46a4767-2e61-4843-806b-4ac0c0d14d4e-catalog-content\") pod \"d46a4767-2e61-4843-806b-4ac0c0d14d4e\" (UID: \"d46a4767-2e61-4843-806b-4ac0c0d14d4e\") " Feb 18 15:21:58 crc kubenswrapper[4896]: I0218 15:21:58.249751 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d46a4767-2e61-4843-806b-4ac0c0d14d4e-utilities" (OuterVolumeSpecName: "utilities") pod "d46a4767-2e61-4843-806b-4ac0c0d14d4e" (UID: "d46a4767-2e61-4843-806b-4ac0c0d14d4e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:21:58 crc kubenswrapper[4896]: I0218 15:21:58.262475 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d46a4767-2e61-4843-806b-4ac0c0d14d4e-kube-api-access-qd46p" (OuterVolumeSpecName: "kube-api-access-qd46p") pod "d46a4767-2e61-4843-806b-4ac0c0d14d4e" (UID: "d46a4767-2e61-4843-806b-4ac0c0d14d4e"). InnerVolumeSpecName "kube-api-access-qd46p". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:21:58 crc kubenswrapper[4896]: I0218 15:21:58.350537 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qd46p\" (UniqueName: \"kubernetes.io/projected/d46a4767-2e61-4843-806b-4ac0c0d14d4e-kube-api-access-qd46p\") on node \"crc\" DevicePath \"\"" Feb 18 15:21:58 crc kubenswrapper[4896]: I0218 15:21:58.350569 4896 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d46a4767-2e61-4843-806b-4ac0c0d14d4e-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 15:21:58 crc kubenswrapper[4896]: I0218 15:21:58.699840 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d46a4767-2e61-4843-806b-4ac0c0d14d4e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d46a4767-2e61-4843-806b-4ac0c0d14d4e" (UID: "d46a4767-2e61-4843-806b-4ac0c0d14d4e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:21:58 crc kubenswrapper[4896]: I0218 15:21:58.755774 4896 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d46a4767-2e61-4843-806b-4ac0c0d14d4e-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 15:21:58 crc kubenswrapper[4896]: I0218 15:21:58.789736 4896 generic.go:334] "Generic (PLEG): container finished" podID="d46a4767-2e61-4843-806b-4ac0c0d14d4e" containerID="69aa7283c33ce19724aae11be6050d24b29018149303c0d40252027b6165c3ed" exitCode=0 Feb 18 15:21:58 crc kubenswrapper[4896]: I0218 15:21:58.789794 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wlxrx" event={"ID":"d46a4767-2e61-4843-806b-4ac0c0d14d4e","Type":"ContainerDied","Data":"69aa7283c33ce19724aae11be6050d24b29018149303c0d40252027b6165c3ed"} Feb 18 15:21:58 crc kubenswrapper[4896]: I0218 15:21:58.789822 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wlxrx" event={"ID":"d46a4767-2e61-4843-806b-4ac0c0d14d4e","Type":"ContainerDied","Data":"6fca9e5f75c0abe0ed9dbd9f627ad8ce47bcc35a807ca2d4a824b5ba8786f1f0"} Feb 18 15:21:58 crc kubenswrapper[4896]: I0218 15:21:58.789845 4896 scope.go:117] "RemoveContainer" containerID="69aa7283c33ce19724aae11be6050d24b29018149303c0d40252027b6165c3ed" Feb 18 15:21:58 crc kubenswrapper[4896]: I0218 15:21:58.789992 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wlxrx" Feb 18 15:21:58 crc kubenswrapper[4896]: I0218 15:21:58.816475 4896 scope.go:117] "RemoveContainer" containerID="e66ecb62c4321d2e4826c3d0ce3462595362e06bdb64068398a857fb137c5c1f" Feb 18 15:21:58 crc kubenswrapper[4896]: I0218 15:21:58.829401 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wlxrx"] Feb 18 15:21:58 crc kubenswrapper[4896]: I0218 15:21:58.836336 4896 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wlxrx"] Feb 18 15:21:58 crc kubenswrapper[4896]: I0218 15:21:58.841267 4896 scope.go:117] "RemoveContainer" containerID="71c0925514e8db0fccce9b34a073fb43c41fac470f44f964e2af36775e772252" Feb 18 15:21:58 crc kubenswrapper[4896]: I0218 15:21:58.860438 4896 scope.go:117] "RemoveContainer" containerID="69aa7283c33ce19724aae11be6050d24b29018149303c0d40252027b6165c3ed" Feb 18 15:21:58 crc kubenswrapper[4896]: E0218 15:21:58.860815 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69aa7283c33ce19724aae11be6050d24b29018149303c0d40252027b6165c3ed\": container with ID starting with 69aa7283c33ce19724aae11be6050d24b29018149303c0d40252027b6165c3ed not found: ID does not exist" containerID="69aa7283c33ce19724aae11be6050d24b29018149303c0d40252027b6165c3ed" Feb 18 15:21:58 crc kubenswrapper[4896]: I0218 15:21:58.860890 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69aa7283c33ce19724aae11be6050d24b29018149303c0d40252027b6165c3ed"} err="failed to get container status \"69aa7283c33ce19724aae11be6050d24b29018149303c0d40252027b6165c3ed\": rpc error: code = NotFound desc = could not find container \"69aa7283c33ce19724aae11be6050d24b29018149303c0d40252027b6165c3ed\": container with ID starting with 69aa7283c33ce19724aae11be6050d24b29018149303c0d40252027b6165c3ed not found: ID does not exist" Feb 18 15:21:58 crc kubenswrapper[4896]: I0218 15:21:58.860927 4896 scope.go:117] "RemoveContainer" containerID="e66ecb62c4321d2e4826c3d0ce3462595362e06bdb64068398a857fb137c5c1f" Feb 18 15:21:58 crc kubenswrapper[4896]: E0218 15:21:58.861366 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e66ecb62c4321d2e4826c3d0ce3462595362e06bdb64068398a857fb137c5c1f\": container with ID starting with e66ecb62c4321d2e4826c3d0ce3462595362e06bdb64068398a857fb137c5c1f not found: ID does not exist" containerID="e66ecb62c4321d2e4826c3d0ce3462595362e06bdb64068398a857fb137c5c1f" Feb 18 15:21:58 crc kubenswrapper[4896]: I0218 15:21:58.861448 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e66ecb62c4321d2e4826c3d0ce3462595362e06bdb64068398a857fb137c5c1f"} err="failed to get container status \"e66ecb62c4321d2e4826c3d0ce3462595362e06bdb64068398a857fb137c5c1f\": rpc error: code = NotFound desc = could not find container \"e66ecb62c4321d2e4826c3d0ce3462595362e06bdb64068398a857fb137c5c1f\": container with ID starting with e66ecb62c4321d2e4826c3d0ce3462595362e06bdb64068398a857fb137c5c1f not found: ID does not exist" Feb 18 15:21:58 crc kubenswrapper[4896]: I0218 15:21:58.861501 4896 scope.go:117] "RemoveContainer" containerID="71c0925514e8db0fccce9b34a073fb43c41fac470f44f964e2af36775e772252" Feb 18 15:21:58 crc kubenswrapper[4896]: E0218 15:21:58.861849 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71c0925514e8db0fccce9b34a073fb43c41fac470f44f964e2af36775e772252\": container with ID starting with 71c0925514e8db0fccce9b34a073fb43c41fac470f44f964e2af36775e772252 not found: ID does not exist" containerID="71c0925514e8db0fccce9b34a073fb43c41fac470f44f964e2af36775e772252" Feb 18 15:21:58 crc kubenswrapper[4896]: I0218 15:21:58.861888 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71c0925514e8db0fccce9b34a073fb43c41fac470f44f964e2af36775e772252"} err="failed to get container status \"71c0925514e8db0fccce9b34a073fb43c41fac470f44f964e2af36775e772252\": rpc error: code = NotFound desc = could not find container \"71c0925514e8db0fccce9b34a073fb43c41fac470f44f964e2af36775e772252\": container with ID starting with 71c0925514e8db0fccce9b34a073fb43c41fac470f44f964e2af36775e772252 not found: ID does not exist" Feb 18 15:21:59 crc kubenswrapper[4896]: I0218 15:21:59.940691 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d46a4767-2e61-4843-806b-4ac0c0d14d4e" path="/var/lib/kubelet/pods/d46a4767-2e61-4843-806b-4ac0c0d14d4e/volumes" Feb 18 15:22:02 crc kubenswrapper[4896]: I0218 15:22:02.480633 4896 scope.go:117] "RemoveContainer" containerID="a3f0a1d2ff900601ef552bacc5623df47e20fe7663f70facd77aa511ce523a4f" Feb 18 15:22:02 crc kubenswrapper[4896]: I0218 15:22:02.499661 4896 scope.go:117] "RemoveContainer" containerID="00f6a64d269d0990c6c301cf52110604d3bb223484458443ba64902717afae5b" Feb 18 15:22:02 crc kubenswrapper[4896]: I0218 15:22:02.541488 4896 scope.go:117] "RemoveContainer" containerID="27c80acd51b5f76959220e73178654655b4ae5f00a7245e3859fbac6aa5c83c8" Feb 18 15:22:02 crc kubenswrapper[4896]: I0218 15:22:02.557962 4896 scope.go:117] "RemoveContainer" containerID="844eb879befbbb02991934eb32d29f0c96ab9a51a086248bf69a58c61f9b2d7a" Feb 18 15:22:03 crc kubenswrapper[4896]: I0218 15:22:03.158377 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7_314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1/util/0.log" Feb 18 15:22:03 crc kubenswrapper[4896]: I0218 15:22:03.358827 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7_314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1/pull/0.log" Feb 18 15:22:03 crc kubenswrapper[4896]: I0218 15:22:03.365995 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7_314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1/pull/0.log" Feb 18 15:22:03 crc kubenswrapper[4896]: I0218 15:22:03.385448 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7_314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1/util/0.log" Feb 18 15:22:03 crc kubenswrapper[4896]: I0218 15:22:03.547687 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7_314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1/util/0.log" Feb 18 15:22:03 crc kubenswrapper[4896]: I0218 15:22:03.553414 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7_314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1/pull/0.log" Feb 18 15:22:03 crc kubenswrapper[4896]: I0218 15:22:03.590589 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a9b3ed1fe9273b725119dcfb777257f08e39bbefccdf592dce2d0dc213gdjm7_314b20e5-bd51-4ffe-9bb0-dd2521ecfcb1/extract/0.log" Feb 18 15:22:03 crc kubenswrapper[4896]: I0218 15:22:03.743163 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gds9w_13955048-63bd-4d03-8aa7-f05d5f69b4e2/extract-utilities/0.log" Feb 18 15:22:03 crc kubenswrapper[4896]: I0218 15:22:03.887308 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gds9w_13955048-63bd-4d03-8aa7-f05d5f69b4e2/extract-utilities/0.log" Feb 18 15:22:03 crc kubenswrapper[4896]: I0218 15:22:03.888126 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gds9w_13955048-63bd-4d03-8aa7-f05d5f69b4e2/extract-content/0.log" Feb 18 15:22:03 crc kubenswrapper[4896]: I0218 15:22:03.917354 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gds9w_13955048-63bd-4d03-8aa7-f05d5f69b4e2/extract-content/0.log" Feb 18 15:22:03 crc kubenswrapper[4896]: I0218 15:22:03.928770 4896 scope.go:117] "RemoveContainer" containerID="7cbf684d610fed53f6fdd655adb1877f0df69baa4d3a58b561f973625a82e83c" Feb 18 15:22:03 crc kubenswrapper[4896]: E0218 15:22:03.929004 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-24rq5_openshift-machine-config-operator(8503734a-8111-4fab-beda-1f0f3b59615b)\"" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" Feb 18 15:22:04 crc kubenswrapper[4896]: I0218 15:22:04.069624 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gds9w_13955048-63bd-4d03-8aa7-f05d5f69b4e2/extract-content/0.log" Feb 18 15:22:04 crc kubenswrapper[4896]: I0218 15:22:04.121602 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gds9w_13955048-63bd-4d03-8aa7-f05d5f69b4e2/extract-utilities/0.log" Feb 18 15:22:04 crc kubenswrapper[4896]: I0218 15:22:04.323149 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fdrnw_354f5a56-d8d3-4fd8-bf60-ee5287896e6c/extract-utilities/0.log" Feb 18 15:22:04 crc kubenswrapper[4896]: I0218 15:22:04.445380 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-gds9w_13955048-63bd-4d03-8aa7-f05d5f69b4e2/registry-server/0.log" Feb 18 15:22:04 crc kubenswrapper[4896]: I0218 15:22:04.578464 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fdrnw_354f5a56-d8d3-4fd8-bf60-ee5287896e6c/extract-content/0.log" Feb 18 15:22:04 crc kubenswrapper[4896]: I0218 15:22:04.578658 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fdrnw_354f5a56-d8d3-4fd8-bf60-ee5287896e6c/extract-content/0.log" Feb 18 15:22:04 crc kubenswrapper[4896]: I0218 15:22:04.585086 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fdrnw_354f5a56-d8d3-4fd8-bf60-ee5287896e6c/extract-utilities/0.log" Feb 18 15:22:04 crc kubenswrapper[4896]: I0218 15:22:04.734986 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fdrnw_354f5a56-d8d3-4fd8-bf60-ee5287896e6c/extract-utilities/0.log" Feb 18 15:22:04 crc kubenswrapper[4896]: I0218 15:22:04.764982 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fdrnw_354f5a56-d8d3-4fd8-bf60-ee5287896e6c/extract-content/0.log" Feb 18 15:22:04 crc kubenswrapper[4896]: I0218 15:22:04.996228 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-5s2vh_18d24b18-a5d0-4fc8-899f-02bf11c3e16c/marketplace-operator/0.log" Feb 18 15:22:05 crc kubenswrapper[4896]: I0218 15:22:05.121024 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-czqrg_8cd2aec2-d292-4074-80db-b06de2ce3118/extract-utilities/0.log" Feb 18 15:22:05 crc kubenswrapper[4896]: I0218 15:22:05.280905 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-fdrnw_354f5a56-d8d3-4fd8-bf60-ee5287896e6c/registry-server/0.log" Feb 18 15:22:05 crc kubenswrapper[4896]: I0218 15:22:05.316760 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-czqrg_8cd2aec2-d292-4074-80db-b06de2ce3118/extract-utilities/0.log" Feb 18 15:22:05 crc kubenswrapper[4896]: I0218 15:22:05.357441 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-czqrg_8cd2aec2-d292-4074-80db-b06de2ce3118/extract-content/0.log" Feb 18 15:22:05 crc kubenswrapper[4896]: I0218 15:22:05.390173 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-czqrg_8cd2aec2-d292-4074-80db-b06de2ce3118/extract-content/0.log" Feb 18 15:22:05 crc kubenswrapper[4896]: I0218 15:22:05.533538 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-czqrg_8cd2aec2-d292-4074-80db-b06de2ce3118/extract-utilities/0.log" Feb 18 15:22:05 crc kubenswrapper[4896]: I0218 15:22:05.554501 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-czqrg_8cd2aec2-d292-4074-80db-b06de2ce3118/extract-content/0.log" Feb 18 15:22:05 crc kubenswrapper[4896]: I0218 15:22:05.639133 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-czqrg_8cd2aec2-d292-4074-80db-b06de2ce3118/registry-server/0.log" Feb 18 15:22:05 crc kubenswrapper[4896]: I0218 15:22:05.767094 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz6dd_bae88343-fc19-473e-9f08-7be40a0ac90d/extract-utilities/0.log" Feb 18 15:22:05 crc kubenswrapper[4896]: I0218 15:22:05.946281 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz6dd_bae88343-fc19-473e-9f08-7be40a0ac90d/extract-content/0.log" Feb 18 15:22:05 crc kubenswrapper[4896]: I0218 15:22:05.965631 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz6dd_bae88343-fc19-473e-9f08-7be40a0ac90d/extract-utilities/0.log" Feb 18 15:22:05 crc kubenswrapper[4896]: I0218 15:22:05.983495 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz6dd_bae88343-fc19-473e-9f08-7be40a0ac90d/extract-content/0.log" Feb 18 15:22:06 crc kubenswrapper[4896]: I0218 15:22:06.139291 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz6dd_bae88343-fc19-473e-9f08-7be40a0ac90d/extract-utilities/0.log" Feb 18 15:22:06 crc kubenswrapper[4896]: I0218 15:22:06.158376 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz6dd_bae88343-fc19-473e-9f08-7be40a0ac90d/extract-content/0.log" Feb 18 15:22:06 crc kubenswrapper[4896]: I0218 15:22:06.590377 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-qz6dd_bae88343-fc19-473e-9f08-7be40a0ac90d/registry-server/0.log" Feb 18 15:22:18 crc kubenswrapper[4896]: I0218 15:22:18.928013 4896 scope.go:117] "RemoveContainer" containerID="7cbf684d610fed53f6fdd655adb1877f0df69baa4d3a58b561f973625a82e83c" Feb 18 15:22:18 crc kubenswrapper[4896]: E0218 15:22:18.929032 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-24rq5_openshift-machine-config-operator(8503734a-8111-4fab-beda-1f0f3b59615b)\"" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" Feb 18 15:22:33 crc kubenswrapper[4896]: I0218 15:22:33.928787 4896 scope.go:117] "RemoveContainer" containerID="7cbf684d610fed53f6fdd655adb1877f0df69baa4d3a58b561f973625a82e83c" Feb 18 15:22:33 crc kubenswrapper[4896]: E0218 15:22:33.929530 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-24rq5_openshift-machine-config-operator(8503734a-8111-4fab-beda-1f0f3b59615b)\"" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" Feb 18 15:22:47 crc kubenswrapper[4896]: I0218 15:22:47.928225 4896 scope.go:117] "RemoveContainer" containerID="7cbf684d610fed53f6fdd655adb1877f0df69baa4d3a58b561f973625a82e83c" Feb 18 15:22:47 crc kubenswrapper[4896]: E0218 15:22:47.929058 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-24rq5_openshift-machine-config-operator(8503734a-8111-4fab-beda-1f0f3b59615b)\"" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" Feb 18 15:22:58 crc kubenswrapper[4896]: I0218 15:22:58.927656 4896 scope.go:117] "RemoveContainer" containerID="7cbf684d610fed53f6fdd655adb1877f0df69baa4d3a58b561f973625a82e83c" Feb 18 15:22:58 crc kubenswrapper[4896]: E0218 15:22:58.928484 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-24rq5_openshift-machine-config-operator(8503734a-8111-4fab-beda-1f0f3b59615b)\"" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" Feb 18 15:23:11 crc kubenswrapper[4896]: I0218 15:23:11.931362 4896 scope.go:117] "RemoveContainer" containerID="7cbf684d610fed53f6fdd655adb1877f0df69baa4d3a58b561f973625a82e83c" Feb 18 15:23:11 crc kubenswrapper[4896]: E0218 15:23:11.932085 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-24rq5_openshift-machine-config-operator(8503734a-8111-4fab-beda-1f0f3b59615b)\"" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" Feb 18 15:23:24 crc kubenswrapper[4896]: I0218 15:23:24.410870 4896 generic.go:334] "Generic (PLEG): container finished" podID="7288ac23-5347-46d8-b5fc-ba214158bf26" containerID="b89b86a45df21b5b2354f3b379c54efaf60c2dd0114aa863916974f0e297a6fd" exitCode=0 Feb 18 15:23:24 crc kubenswrapper[4896]: I0218 15:23:24.410961 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-86dmm/must-gather-7jbtf" event={"ID":"7288ac23-5347-46d8-b5fc-ba214158bf26","Type":"ContainerDied","Data":"b89b86a45df21b5b2354f3b379c54efaf60c2dd0114aa863916974f0e297a6fd"} Feb 18 15:23:24 crc kubenswrapper[4896]: I0218 15:23:24.412050 4896 scope.go:117] "RemoveContainer" containerID="b89b86a45df21b5b2354f3b379c54efaf60c2dd0114aa863916974f0e297a6fd" Feb 18 15:23:24 crc kubenswrapper[4896]: I0218 15:23:24.699431 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-86dmm_must-gather-7jbtf_7288ac23-5347-46d8-b5fc-ba214158bf26/gather/0.log" Feb 18 15:23:24 crc kubenswrapper[4896]: I0218 15:23:24.927837 4896 scope.go:117] "RemoveContainer" containerID="7cbf684d610fed53f6fdd655adb1877f0df69baa4d3a58b561f973625a82e83c" Feb 18 15:23:24 crc kubenswrapper[4896]: E0218 15:23:24.928044 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-24rq5_openshift-machine-config-operator(8503734a-8111-4fab-beda-1f0f3b59615b)\"" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" Feb 18 15:23:31 crc kubenswrapper[4896]: I0218 15:23:31.763318 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-86dmm/must-gather-7jbtf"] Feb 18 15:23:31 crc kubenswrapper[4896]: I0218 15:23:31.764062 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-86dmm/must-gather-7jbtf" podUID="7288ac23-5347-46d8-b5fc-ba214158bf26" containerName="copy" containerID="cri-o://22108ab6b903f1f8329b2fb04c9dba418c8e363018e0dbd2e709907b76d8bee9" gracePeriod=2 Feb 18 15:23:31 crc kubenswrapper[4896]: I0218 15:23:31.768028 4896 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-86dmm/must-gather-7jbtf"] Feb 18 15:23:32 crc kubenswrapper[4896]: I0218 15:23:32.163222 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-86dmm_must-gather-7jbtf_7288ac23-5347-46d8-b5fc-ba214158bf26/copy/0.log" Feb 18 15:23:32 crc kubenswrapper[4896]: I0218 15:23:32.164189 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-86dmm/must-gather-7jbtf" Feb 18 15:23:32 crc kubenswrapper[4896]: I0218 15:23:32.361448 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5hbz\" (UniqueName: \"kubernetes.io/projected/7288ac23-5347-46d8-b5fc-ba214158bf26-kube-api-access-q5hbz\") pod \"7288ac23-5347-46d8-b5fc-ba214158bf26\" (UID: \"7288ac23-5347-46d8-b5fc-ba214158bf26\") " Feb 18 15:23:32 crc kubenswrapper[4896]: I0218 15:23:32.361518 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7288ac23-5347-46d8-b5fc-ba214158bf26-must-gather-output\") pod \"7288ac23-5347-46d8-b5fc-ba214158bf26\" (UID: \"7288ac23-5347-46d8-b5fc-ba214158bf26\") " Feb 18 15:23:32 crc kubenswrapper[4896]: I0218 15:23:32.369426 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7288ac23-5347-46d8-b5fc-ba214158bf26-kube-api-access-q5hbz" (OuterVolumeSpecName: "kube-api-access-q5hbz") pod "7288ac23-5347-46d8-b5fc-ba214158bf26" (UID: "7288ac23-5347-46d8-b5fc-ba214158bf26"). InnerVolumeSpecName "kube-api-access-q5hbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:23:32 crc kubenswrapper[4896]: I0218 15:23:32.436643 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7288ac23-5347-46d8-b5fc-ba214158bf26-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "7288ac23-5347-46d8-b5fc-ba214158bf26" (UID: "7288ac23-5347-46d8-b5fc-ba214158bf26"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:23:32 crc kubenswrapper[4896]: I0218 15:23:32.463152 4896 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/7288ac23-5347-46d8-b5fc-ba214158bf26-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 18 15:23:32 crc kubenswrapper[4896]: I0218 15:23:32.463188 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5hbz\" (UniqueName: \"kubernetes.io/projected/7288ac23-5347-46d8-b5fc-ba214158bf26-kube-api-access-q5hbz\") on node \"crc\" DevicePath \"\"" Feb 18 15:23:32 crc kubenswrapper[4896]: I0218 15:23:32.466676 4896 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-86dmm_must-gather-7jbtf_7288ac23-5347-46d8-b5fc-ba214158bf26/copy/0.log" Feb 18 15:23:32 crc kubenswrapper[4896]: I0218 15:23:32.467004 4896 generic.go:334] "Generic (PLEG): container finished" podID="7288ac23-5347-46d8-b5fc-ba214158bf26" containerID="22108ab6b903f1f8329b2fb04c9dba418c8e363018e0dbd2e709907b76d8bee9" exitCode=143 Feb 18 15:23:32 crc kubenswrapper[4896]: I0218 15:23:32.467036 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-86dmm/must-gather-7jbtf" Feb 18 15:23:32 crc kubenswrapper[4896]: I0218 15:23:32.467053 4896 scope.go:117] "RemoveContainer" containerID="22108ab6b903f1f8329b2fb04c9dba418c8e363018e0dbd2e709907b76d8bee9" Feb 18 15:23:32 crc kubenswrapper[4896]: I0218 15:23:32.510752 4896 scope.go:117] "RemoveContainer" containerID="b89b86a45df21b5b2354f3b379c54efaf60c2dd0114aa863916974f0e297a6fd" Feb 18 15:23:32 crc kubenswrapper[4896]: I0218 15:23:32.560159 4896 scope.go:117] "RemoveContainer" containerID="22108ab6b903f1f8329b2fb04c9dba418c8e363018e0dbd2e709907b76d8bee9" Feb 18 15:23:32 crc kubenswrapper[4896]: E0218 15:23:32.561649 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22108ab6b903f1f8329b2fb04c9dba418c8e363018e0dbd2e709907b76d8bee9\": container with ID starting with 22108ab6b903f1f8329b2fb04c9dba418c8e363018e0dbd2e709907b76d8bee9 not found: ID does not exist" containerID="22108ab6b903f1f8329b2fb04c9dba418c8e363018e0dbd2e709907b76d8bee9" Feb 18 15:23:32 crc kubenswrapper[4896]: I0218 15:23:32.561685 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22108ab6b903f1f8329b2fb04c9dba418c8e363018e0dbd2e709907b76d8bee9"} err="failed to get container status \"22108ab6b903f1f8329b2fb04c9dba418c8e363018e0dbd2e709907b76d8bee9\": rpc error: code = NotFound desc = could not find container \"22108ab6b903f1f8329b2fb04c9dba418c8e363018e0dbd2e709907b76d8bee9\": container with ID starting with 22108ab6b903f1f8329b2fb04c9dba418c8e363018e0dbd2e709907b76d8bee9 not found: ID does not exist" Feb 18 15:23:32 crc kubenswrapper[4896]: I0218 15:23:32.561713 4896 scope.go:117] "RemoveContainer" containerID="b89b86a45df21b5b2354f3b379c54efaf60c2dd0114aa863916974f0e297a6fd" Feb 18 15:23:32 crc kubenswrapper[4896]: E0218 15:23:32.565564 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b89b86a45df21b5b2354f3b379c54efaf60c2dd0114aa863916974f0e297a6fd\": container with ID starting with b89b86a45df21b5b2354f3b379c54efaf60c2dd0114aa863916974f0e297a6fd not found: ID does not exist" containerID="b89b86a45df21b5b2354f3b379c54efaf60c2dd0114aa863916974f0e297a6fd" Feb 18 15:23:32 crc kubenswrapper[4896]: I0218 15:23:32.565613 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b89b86a45df21b5b2354f3b379c54efaf60c2dd0114aa863916974f0e297a6fd"} err="failed to get container status \"b89b86a45df21b5b2354f3b379c54efaf60c2dd0114aa863916974f0e297a6fd\": rpc error: code = NotFound desc = could not find container \"b89b86a45df21b5b2354f3b379c54efaf60c2dd0114aa863916974f0e297a6fd\": container with ID starting with b89b86a45df21b5b2354f3b379c54efaf60c2dd0114aa863916974f0e297a6fd not found: ID does not exist" Feb 18 15:23:33 crc kubenswrapper[4896]: I0218 15:23:33.938901 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7288ac23-5347-46d8-b5fc-ba214158bf26" path="/var/lib/kubelet/pods/7288ac23-5347-46d8-b5fc-ba214158bf26/volumes" Feb 18 15:23:35 crc kubenswrapper[4896]: I0218 15:23:35.927977 4896 scope.go:117] "RemoveContainer" containerID="7cbf684d610fed53f6fdd655adb1877f0df69baa4d3a58b561f973625a82e83c" Feb 18 15:23:35 crc kubenswrapper[4896]: E0218 15:23:35.928701 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-24rq5_openshift-machine-config-operator(8503734a-8111-4fab-beda-1f0f3b59615b)\"" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" Feb 18 15:23:50 crc kubenswrapper[4896]: I0218 15:23:50.927855 4896 scope.go:117] "RemoveContainer" containerID="7cbf684d610fed53f6fdd655adb1877f0df69baa4d3a58b561f973625a82e83c" Feb 18 15:23:50 crc kubenswrapper[4896]: E0218 15:23:50.928548 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-24rq5_openshift-machine-config-operator(8503734a-8111-4fab-beda-1f0f3b59615b)\"" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" Feb 18 15:24:02 crc kubenswrapper[4896]: I0218 15:24:02.928299 4896 scope.go:117] "RemoveContainer" containerID="7cbf684d610fed53f6fdd655adb1877f0df69baa4d3a58b561f973625a82e83c" Feb 18 15:24:02 crc kubenswrapper[4896]: E0218 15:24:02.928828 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-24rq5_openshift-machine-config-operator(8503734a-8111-4fab-beda-1f0f3b59615b)\"" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" Feb 18 15:24:13 crc kubenswrapper[4896]: I0218 15:24:13.929330 4896 scope.go:117] "RemoveContainer" containerID="7cbf684d610fed53f6fdd655adb1877f0df69baa4d3a58b561f973625a82e83c" Feb 18 15:24:13 crc kubenswrapper[4896]: E0218 15:24:13.929891 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-24rq5_openshift-machine-config-operator(8503734a-8111-4fab-beda-1f0f3b59615b)\"" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" Feb 18 15:24:24 crc kubenswrapper[4896]: I0218 15:24:24.927624 4896 scope.go:117] "RemoveContainer" containerID="7cbf684d610fed53f6fdd655adb1877f0df69baa4d3a58b561f973625a82e83c" Feb 18 15:24:24 crc kubenswrapper[4896]: E0218 15:24:24.928330 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-24rq5_openshift-machine-config-operator(8503734a-8111-4fab-beda-1f0f3b59615b)\"" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" Feb 18 15:24:39 crc kubenswrapper[4896]: I0218 15:24:39.933654 4896 scope.go:117] "RemoveContainer" containerID="7cbf684d610fed53f6fdd655adb1877f0df69baa4d3a58b561f973625a82e83c" Feb 18 15:24:39 crc kubenswrapper[4896]: E0218 15:24:39.934284 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-24rq5_openshift-machine-config-operator(8503734a-8111-4fab-beda-1f0f3b59615b)\"" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" Feb 18 15:24:53 crc kubenswrapper[4896]: I0218 15:24:53.928406 4896 scope.go:117] "RemoveContainer" containerID="7cbf684d610fed53f6fdd655adb1877f0df69baa4d3a58b561f973625a82e83c" Feb 18 15:24:53 crc kubenswrapper[4896]: E0218 15:24:53.929409 4896 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-24rq5_openshift-machine-config-operator(8503734a-8111-4fab-beda-1f0f3b59615b)\"" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" Feb 18 15:25:04 crc kubenswrapper[4896]: I0218 15:25:04.407675 4896 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-v5c6v"] Feb 18 15:25:04 crc kubenswrapper[4896]: E0218 15:25:04.408464 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7288ac23-5347-46d8-b5fc-ba214158bf26" containerName="copy" Feb 18 15:25:04 crc kubenswrapper[4896]: I0218 15:25:04.408480 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="7288ac23-5347-46d8-b5fc-ba214158bf26" containerName="copy" Feb 18 15:25:04 crc kubenswrapper[4896]: E0218 15:25:04.408501 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d46a4767-2e61-4843-806b-4ac0c0d14d4e" containerName="extract-utilities" Feb 18 15:25:04 crc kubenswrapper[4896]: I0218 15:25:04.408509 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="d46a4767-2e61-4843-806b-4ac0c0d14d4e" containerName="extract-utilities" Feb 18 15:25:04 crc kubenswrapper[4896]: E0218 15:25:04.408526 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7288ac23-5347-46d8-b5fc-ba214158bf26" containerName="gather" Feb 18 15:25:04 crc kubenswrapper[4896]: I0218 15:25:04.408533 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="7288ac23-5347-46d8-b5fc-ba214158bf26" containerName="gather" Feb 18 15:25:04 crc kubenswrapper[4896]: E0218 15:25:04.408544 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d46a4767-2e61-4843-806b-4ac0c0d14d4e" containerName="registry-server" Feb 18 15:25:04 crc kubenswrapper[4896]: I0218 15:25:04.408551 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="d46a4767-2e61-4843-806b-4ac0c0d14d4e" containerName="registry-server" Feb 18 15:25:04 crc kubenswrapper[4896]: E0218 15:25:04.408562 4896 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d46a4767-2e61-4843-806b-4ac0c0d14d4e" containerName="extract-content" Feb 18 15:25:04 crc kubenswrapper[4896]: I0218 15:25:04.408570 4896 state_mem.go:107] "Deleted CPUSet assignment" podUID="d46a4767-2e61-4843-806b-4ac0c0d14d4e" containerName="extract-content" Feb 18 15:25:04 crc kubenswrapper[4896]: I0218 15:25:04.408717 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="d46a4767-2e61-4843-806b-4ac0c0d14d4e" containerName="registry-server" Feb 18 15:25:04 crc kubenswrapper[4896]: I0218 15:25:04.408732 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="7288ac23-5347-46d8-b5fc-ba214158bf26" containerName="gather" Feb 18 15:25:04 crc kubenswrapper[4896]: I0218 15:25:04.408750 4896 memory_manager.go:354] "RemoveStaleState removing state" podUID="7288ac23-5347-46d8-b5fc-ba214158bf26" containerName="copy" Feb 18 15:25:04 crc kubenswrapper[4896]: I0218 15:25:04.409709 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v5c6v" Feb 18 15:25:04 crc kubenswrapper[4896]: I0218 15:25:04.429502 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v5c6v"] Feb 18 15:25:04 crc kubenswrapper[4896]: I0218 15:25:04.526317 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03c189c0-d5fb-4cbc-9e38-fcaf54263a35-utilities\") pod \"redhat-marketplace-v5c6v\" (UID: \"03c189c0-d5fb-4cbc-9e38-fcaf54263a35\") " pod="openshift-marketplace/redhat-marketplace-v5c6v" Feb 18 15:25:04 crc kubenswrapper[4896]: I0218 15:25:04.526878 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03c189c0-d5fb-4cbc-9e38-fcaf54263a35-catalog-content\") pod \"redhat-marketplace-v5c6v\" (UID: \"03c189c0-d5fb-4cbc-9e38-fcaf54263a35\") " pod="openshift-marketplace/redhat-marketplace-v5c6v" Feb 18 15:25:04 crc kubenswrapper[4896]: I0218 15:25:04.527042 4896 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxf6b\" (UniqueName: \"kubernetes.io/projected/03c189c0-d5fb-4cbc-9e38-fcaf54263a35-kube-api-access-pxf6b\") pod \"redhat-marketplace-v5c6v\" (UID: \"03c189c0-d5fb-4cbc-9e38-fcaf54263a35\") " pod="openshift-marketplace/redhat-marketplace-v5c6v" Feb 18 15:25:04 crc kubenswrapper[4896]: I0218 15:25:04.627999 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxf6b\" (UniqueName: \"kubernetes.io/projected/03c189c0-d5fb-4cbc-9e38-fcaf54263a35-kube-api-access-pxf6b\") pod \"redhat-marketplace-v5c6v\" (UID: \"03c189c0-d5fb-4cbc-9e38-fcaf54263a35\") " pod="openshift-marketplace/redhat-marketplace-v5c6v" Feb 18 15:25:04 crc kubenswrapper[4896]: I0218 15:25:04.628085 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03c189c0-d5fb-4cbc-9e38-fcaf54263a35-utilities\") pod \"redhat-marketplace-v5c6v\" (UID: \"03c189c0-d5fb-4cbc-9e38-fcaf54263a35\") " pod="openshift-marketplace/redhat-marketplace-v5c6v" Feb 18 15:25:04 crc kubenswrapper[4896]: I0218 15:25:04.628118 4896 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03c189c0-d5fb-4cbc-9e38-fcaf54263a35-catalog-content\") pod \"redhat-marketplace-v5c6v\" (UID: \"03c189c0-d5fb-4cbc-9e38-fcaf54263a35\") " pod="openshift-marketplace/redhat-marketplace-v5c6v" Feb 18 15:25:04 crc kubenswrapper[4896]: I0218 15:25:04.628644 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03c189c0-d5fb-4cbc-9e38-fcaf54263a35-utilities\") pod \"redhat-marketplace-v5c6v\" (UID: \"03c189c0-d5fb-4cbc-9e38-fcaf54263a35\") " pod="openshift-marketplace/redhat-marketplace-v5c6v" Feb 18 15:25:04 crc kubenswrapper[4896]: I0218 15:25:04.628676 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03c189c0-d5fb-4cbc-9e38-fcaf54263a35-catalog-content\") pod \"redhat-marketplace-v5c6v\" (UID: \"03c189c0-d5fb-4cbc-9e38-fcaf54263a35\") " pod="openshift-marketplace/redhat-marketplace-v5c6v" Feb 18 15:25:04 crc kubenswrapper[4896]: I0218 15:25:04.649029 4896 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxf6b\" (UniqueName: \"kubernetes.io/projected/03c189c0-d5fb-4cbc-9e38-fcaf54263a35-kube-api-access-pxf6b\") pod \"redhat-marketplace-v5c6v\" (UID: \"03c189c0-d5fb-4cbc-9e38-fcaf54263a35\") " pod="openshift-marketplace/redhat-marketplace-v5c6v" Feb 18 15:25:04 crc kubenswrapper[4896]: I0218 15:25:04.729092 4896 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v5c6v" Feb 18 15:25:04 crc kubenswrapper[4896]: I0218 15:25:04.928455 4896 scope.go:117] "RemoveContainer" containerID="7cbf684d610fed53f6fdd655adb1877f0df69baa4d3a58b561f973625a82e83c" Feb 18 15:25:05 crc kubenswrapper[4896]: I0218 15:25:05.134163 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" event={"ID":"8503734a-8111-4fab-beda-1f0f3b59615b","Type":"ContainerStarted","Data":"8a2626b90566022dacc8a093fa9dc53568a6310c6859fa806f9cbe79b3800686"} Feb 18 15:25:05 crc kubenswrapper[4896]: I0218 15:25:05.178834 4896 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v5c6v"] Feb 18 15:25:05 crc kubenswrapper[4896]: W0218 15:25:05.182528 4896 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03c189c0_d5fb_4cbc_9e38_fcaf54263a35.slice/crio-ff77eb7ea46b24dfda766043f92cee092674ed571490e07edfb5299ba261a3dc WatchSource:0}: Error finding container ff77eb7ea46b24dfda766043f92cee092674ed571490e07edfb5299ba261a3dc: Status 404 returned error can't find the container with id ff77eb7ea46b24dfda766043f92cee092674ed571490e07edfb5299ba261a3dc Feb 18 15:25:06 crc kubenswrapper[4896]: I0218 15:25:06.141319 4896 generic.go:334] "Generic (PLEG): container finished" podID="03c189c0-d5fb-4cbc-9e38-fcaf54263a35" containerID="515a809a609ce67fc0b0ded9b3b09bf3518e3305c5fb0cf2889280cc0a6b0e56" exitCode=0 Feb 18 15:25:06 crc kubenswrapper[4896]: I0218 15:25:06.141469 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5c6v" event={"ID":"03c189c0-d5fb-4cbc-9e38-fcaf54263a35","Type":"ContainerDied","Data":"515a809a609ce67fc0b0ded9b3b09bf3518e3305c5fb0cf2889280cc0a6b0e56"} Feb 18 15:25:06 crc kubenswrapper[4896]: I0218 15:25:06.141874 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5c6v" event={"ID":"03c189c0-d5fb-4cbc-9e38-fcaf54263a35","Type":"ContainerStarted","Data":"ff77eb7ea46b24dfda766043f92cee092674ed571490e07edfb5299ba261a3dc"} Feb 18 15:25:06 crc kubenswrapper[4896]: I0218 15:25:06.143479 4896 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 18 15:25:07 crc kubenswrapper[4896]: I0218 15:25:07.150837 4896 generic.go:334] "Generic (PLEG): container finished" podID="03c189c0-d5fb-4cbc-9e38-fcaf54263a35" containerID="999f98cf09a0734eb5263b0c4402447cb6be3791c34f5d5676c26cd0528cf689" exitCode=0 Feb 18 15:25:07 crc kubenswrapper[4896]: I0218 15:25:07.150905 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5c6v" event={"ID":"03c189c0-d5fb-4cbc-9e38-fcaf54263a35","Type":"ContainerDied","Data":"999f98cf09a0734eb5263b0c4402447cb6be3791c34f5d5676c26cd0528cf689"} Feb 18 15:25:08 crc kubenswrapper[4896]: I0218 15:25:08.159261 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5c6v" event={"ID":"03c189c0-d5fb-4cbc-9e38-fcaf54263a35","Type":"ContainerStarted","Data":"58176d484251a3a1c1d022997e67717644cbc5a0bf8a24a1e0d0a57d4e215683"} Feb 18 15:25:08 crc kubenswrapper[4896]: I0218 15:25:08.178267 4896 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-v5c6v" podStartSLOduration=2.6642398309999997 podStartE2EDuration="4.178248264s" podCreationTimestamp="2026-02-18 15:25:04 +0000 UTC" firstStartedPulling="2026-02-18 15:25:06.143193864 +0000 UTC m=+1805.072528077" lastFinishedPulling="2026-02-18 15:25:07.657202297 +0000 UTC m=+1806.586536510" observedRunningTime="2026-02-18 15:25:08.175460097 +0000 UTC m=+1807.104794310" watchObservedRunningTime="2026-02-18 15:25:08.178248264 +0000 UTC m=+1807.107582477" Feb 18 15:25:14 crc kubenswrapper[4896]: I0218 15:25:14.729871 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-v5c6v" Feb 18 15:25:14 crc kubenswrapper[4896]: I0218 15:25:14.730684 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-v5c6v" Feb 18 15:25:14 crc kubenswrapper[4896]: I0218 15:25:14.769142 4896 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-v5c6v" Feb 18 15:25:15 crc kubenswrapper[4896]: I0218 15:25:15.275183 4896 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-v5c6v" Feb 18 15:25:16 crc kubenswrapper[4896]: I0218 15:25:16.201640 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v5c6v"] Feb 18 15:25:17 crc kubenswrapper[4896]: I0218 15:25:17.225127 4896 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-v5c6v" podUID="03c189c0-d5fb-4cbc-9e38-fcaf54263a35" containerName="registry-server" containerID="cri-o://58176d484251a3a1c1d022997e67717644cbc5a0bf8a24a1e0d0a57d4e215683" gracePeriod=2 Feb 18 15:25:17 crc kubenswrapper[4896]: I0218 15:25:17.687107 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v5c6v" Feb 18 15:25:17 crc kubenswrapper[4896]: I0218 15:25:17.809443 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03c189c0-d5fb-4cbc-9e38-fcaf54263a35-utilities\") pod \"03c189c0-d5fb-4cbc-9e38-fcaf54263a35\" (UID: \"03c189c0-d5fb-4cbc-9e38-fcaf54263a35\") " Feb 18 15:25:17 crc kubenswrapper[4896]: I0218 15:25:17.809534 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxf6b\" (UniqueName: \"kubernetes.io/projected/03c189c0-d5fb-4cbc-9e38-fcaf54263a35-kube-api-access-pxf6b\") pod \"03c189c0-d5fb-4cbc-9e38-fcaf54263a35\" (UID: \"03c189c0-d5fb-4cbc-9e38-fcaf54263a35\") " Feb 18 15:25:17 crc kubenswrapper[4896]: I0218 15:25:17.809619 4896 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03c189c0-d5fb-4cbc-9e38-fcaf54263a35-catalog-content\") pod \"03c189c0-d5fb-4cbc-9e38-fcaf54263a35\" (UID: \"03c189c0-d5fb-4cbc-9e38-fcaf54263a35\") " Feb 18 15:25:17 crc kubenswrapper[4896]: I0218 15:25:17.810805 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03c189c0-d5fb-4cbc-9e38-fcaf54263a35-utilities" (OuterVolumeSpecName: "utilities") pod "03c189c0-d5fb-4cbc-9e38-fcaf54263a35" (UID: "03c189c0-d5fb-4cbc-9e38-fcaf54263a35"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:25:17 crc kubenswrapper[4896]: I0218 15:25:17.815074 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03c189c0-d5fb-4cbc-9e38-fcaf54263a35-kube-api-access-pxf6b" (OuterVolumeSpecName: "kube-api-access-pxf6b") pod "03c189c0-d5fb-4cbc-9e38-fcaf54263a35" (UID: "03c189c0-d5fb-4cbc-9e38-fcaf54263a35"). InnerVolumeSpecName "kube-api-access-pxf6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 18 15:25:17 crc kubenswrapper[4896]: I0218 15:25:17.833556 4896 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03c189c0-d5fb-4cbc-9e38-fcaf54263a35-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "03c189c0-d5fb-4cbc-9e38-fcaf54263a35" (UID: "03c189c0-d5fb-4cbc-9e38-fcaf54263a35"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 18 15:25:17 crc kubenswrapper[4896]: I0218 15:25:17.910892 4896 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03c189c0-d5fb-4cbc-9e38-fcaf54263a35-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:17 crc kubenswrapper[4896]: I0218 15:25:17.910930 4896 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03c189c0-d5fb-4cbc-9e38-fcaf54263a35-utilities\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:17 crc kubenswrapper[4896]: I0218 15:25:17.910940 4896 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxf6b\" (UniqueName: \"kubernetes.io/projected/03c189c0-d5fb-4cbc-9e38-fcaf54263a35-kube-api-access-pxf6b\") on node \"crc\" DevicePath \"\"" Feb 18 15:25:18 crc kubenswrapper[4896]: I0218 15:25:18.231473 4896 generic.go:334] "Generic (PLEG): container finished" podID="03c189c0-d5fb-4cbc-9e38-fcaf54263a35" containerID="58176d484251a3a1c1d022997e67717644cbc5a0bf8a24a1e0d0a57d4e215683" exitCode=0 Feb 18 15:25:18 crc kubenswrapper[4896]: I0218 15:25:18.231515 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5c6v" event={"ID":"03c189c0-d5fb-4cbc-9e38-fcaf54263a35","Type":"ContainerDied","Data":"58176d484251a3a1c1d022997e67717644cbc5a0bf8a24a1e0d0a57d4e215683"} Feb 18 15:25:18 crc kubenswrapper[4896]: I0218 15:25:18.231537 4896 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v5c6v" event={"ID":"03c189c0-d5fb-4cbc-9e38-fcaf54263a35","Type":"ContainerDied","Data":"ff77eb7ea46b24dfda766043f92cee092674ed571490e07edfb5299ba261a3dc"} Feb 18 15:25:18 crc kubenswrapper[4896]: I0218 15:25:18.231553 4896 scope.go:117] "RemoveContainer" containerID="58176d484251a3a1c1d022997e67717644cbc5a0bf8a24a1e0d0a57d4e215683" Feb 18 15:25:18 crc kubenswrapper[4896]: I0218 15:25:18.232182 4896 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v5c6v" Feb 18 15:25:18 crc kubenswrapper[4896]: I0218 15:25:18.249158 4896 scope.go:117] "RemoveContainer" containerID="999f98cf09a0734eb5263b0c4402447cb6be3791c34f5d5676c26cd0528cf689" Feb 18 15:25:18 crc kubenswrapper[4896]: I0218 15:25:18.250600 4896 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-v5c6v"] Feb 18 15:25:18 crc kubenswrapper[4896]: I0218 15:25:18.256355 4896 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-v5c6v"] Feb 18 15:25:18 crc kubenswrapper[4896]: I0218 15:25:18.266792 4896 scope.go:117] "RemoveContainer" containerID="515a809a609ce67fc0b0ded9b3b09bf3518e3305c5fb0cf2889280cc0a6b0e56" Feb 18 15:25:18 crc kubenswrapper[4896]: I0218 15:25:18.286658 4896 scope.go:117] "RemoveContainer" containerID="58176d484251a3a1c1d022997e67717644cbc5a0bf8a24a1e0d0a57d4e215683" Feb 18 15:25:18 crc kubenswrapper[4896]: E0218 15:25:18.287122 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58176d484251a3a1c1d022997e67717644cbc5a0bf8a24a1e0d0a57d4e215683\": container with ID starting with 58176d484251a3a1c1d022997e67717644cbc5a0bf8a24a1e0d0a57d4e215683 not found: ID does not exist" containerID="58176d484251a3a1c1d022997e67717644cbc5a0bf8a24a1e0d0a57d4e215683" Feb 18 15:25:18 crc kubenswrapper[4896]: I0218 15:25:18.287164 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58176d484251a3a1c1d022997e67717644cbc5a0bf8a24a1e0d0a57d4e215683"} err="failed to get container status \"58176d484251a3a1c1d022997e67717644cbc5a0bf8a24a1e0d0a57d4e215683\": rpc error: code = NotFound desc = could not find container \"58176d484251a3a1c1d022997e67717644cbc5a0bf8a24a1e0d0a57d4e215683\": container with ID starting with 58176d484251a3a1c1d022997e67717644cbc5a0bf8a24a1e0d0a57d4e215683 not found: ID does not exist" Feb 18 15:25:18 crc kubenswrapper[4896]: I0218 15:25:18.287195 4896 scope.go:117] "RemoveContainer" containerID="999f98cf09a0734eb5263b0c4402447cb6be3791c34f5d5676c26cd0528cf689" Feb 18 15:25:18 crc kubenswrapper[4896]: E0218 15:25:18.287593 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"999f98cf09a0734eb5263b0c4402447cb6be3791c34f5d5676c26cd0528cf689\": container with ID starting with 999f98cf09a0734eb5263b0c4402447cb6be3791c34f5d5676c26cd0528cf689 not found: ID does not exist" containerID="999f98cf09a0734eb5263b0c4402447cb6be3791c34f5d5676c26cd0528cf689" Feb 18 15:25:18 crc kubenswrapper[4896]: I0218 15:25:18.287643 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"999f98cf09a0734eb5263b0c4402447cb6be3791c34f5d5676c26cd0528cf689"} err="failed to get container status \"999f98cf09a0734eb5263b0c4402447cb6be3791c34f5d5676c26cd0528cf689\": rpc error: code = NotFound desc = could not find container \"999f98cf09a0734eb5263b0c4402447cb6be3791c34f5d5676c26cd0528cf689\": container with ID starting with 999f98cf09a0734eb5263b0c4402447cb6be3791c34f5d5676c26cd0528cf689 not found: ID does not exist" Feb 18 15:25:18 crc kubenswrapper[4896]: I0218 15:25:18.287674 4896 scope.go:117] "RemoveContainer" containerID="515a809a609ce67fc0b0ded9b3b09bf3518e3305c5fb0cf2889280cc0a6b0e56" Feb 18 15:25:18 crc kubenswrapper[4896]: E0218 15:25:18.287950 4896 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"515a809a609ce67fc0b0ded9b3b09bf3518e3305c5fb0cf2889280cc0a6b0e56\": container with ID starting with 515a809a609ce67fc0b0ded9b3b09bf3518e3305c5fb0cf2889280cc0a6b0e56 not found: ID does not exist" containerID="515a809a609ce67fc0b0ded9b3b09bf3518e3305c5fb0cf2889280cc0a6b0e56" Feb 18 15:25:18 crc kubenswrapper[4896]: I0218 15:25:18.287986 4896 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"515a809a609ce67fc0b0ded9b3b09bf3518e3305c5fb0cf2889280cc0a6b0e56"} err="failed to get container status \"515a809a609ce67fc0b0ded9b3b09bf3518e3305c5fb0cf2889280cc0a6b0e56\": rpc error: code = NotFound desc = could not find container \"515a809a609ce67fc0b0ded9b3b09bf3518e3305c5fb0cf2889280cc0a6b0e56\": container with ID starting with 515a809a609ce67fc0b0ded9b3b09bf3518e3305c5fb0cf2889280cc0a6b0e56 not found: ID does not exist" Feb 18 15:25:19 crc kubenswrapper[4896]: I0218 15:25:19.936737 4896 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03c189c0-d5fb-4cbc-9e38-fcaf54263a35" path="/var/lib/kubelet/pods/03c189c0-d5fb-4cbc-9e38-fcaf54263a35/volumes" Feb 18 15:27:34 crc kubenswrapper[4896]: I0218 15:27:34.075185 4896 patch_prober.go:28] interesting pod/machine-config-daemon-24rq5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:27:34 crc kubenswrapper[4896]: I0218 15:27:34.075836 4896 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 18 15:28:04 crc kubenswrapper[4896]: I0218 15:28:04.075508 4896 patch_prober.go:28] interesting pod/machine-config-daemon-24rq5 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 18 15:28:04 crc kubenswrapper[4896]: I0218 15:28:04.076224 4896 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-24rq5" podUID="8503734a-8111-4fab-beda-1f0f3b59615b" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515145355250024452 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015145355251017370 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015145351044016507 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015145351044015457 5ustar corecore